{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/airflow"},"x-facet":{"type":"skill","slug":"airflow","display":"Airflow","count":83},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1bd2d1b2-84f"},"title":"Senior Machine Learning Researcher","description":"<p>We are seeking a senior machine learning researcher to join our Core AI team.</p>\n<p>As part of the team, you will help solve complex business problems by developing viable cutting-edge AI/ML solutions.</p>\n<p>You will develop and implement creative solutions that fundamentally transform business processes, delivering breakthrough improvements rather than incremental changes.</p>\n<p>You will work closely with other AI/ML researchers and engineers, SWEs, product owners/managers, and business stakeholders, and participate in the full lifecycle of solution development, including requirements gathering with business, experimentation and algorithmic exploration, development, and assistance with productization.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Work independently or as part of a team to help design and implement high accuracy and delightful user experience solutions utilizing ML, NLP, GenAI, Agentic technologies.</li>\n</ul>\n<ul>\n<li>Participate in all aspects of solution development, including ideation and requirement gathering with business stakeholders, experimentation and exploration to identify strong solution approaches, solution development, etc.</li>\n</ul>\n<ul>\n<li>Prototype, test, and iterate on novel AI models and approaches to solve complex business challenges.</li>\n</ul>\n<ul>\n<li>Collaborate with cross-functional teams to identify opportunities where AI can create significant business value, and transition solutions into production systems.</li>\n</ul>\n<ul>\n<li>Research and stay updated with the latest advancements in machine learning and AI technologies.</li>\n</ul>\n<ul>\n<li>Participate in code reviews, technical discussions, and knowledge sharing sessions.</li>\n</ul>\n<ul>\n<li>Communicate technical concepts and transformative ideas effectively to both technical and non-technical stakeholders.</li>\n</ul>\n<p>Required Skills &amp; Qualifications:</p>\n<ul>\n<li>Bachelor&#39;s with 10+ years, Master&#39;s with 7+ years, or PhD with 5+ years in Computer Science, Data Science, Machine Learning, or related field.</li>\n</ul>\n<ul>\n<li>Deep expertise and proven ability in developing high accuracy/value solutions to business problems in the NLP, Generative AI, Agentic AI, and/or ML space.</li>\n</ul>\n<ul>\n<li>Hands-on experience with data processing, experimentation, and exploration.</li>\n</ul>\n<ul>\n<li>Strong programming skills in Python.</li>\n</ul>\n<ul>\n<li>Experience with cloud platforms (AWS, Azure, GCP) for deploying ML solutions.</li>\n</ul>\n<ul>\n<li>Excellent problem-solving skills and attention to detail.</li>\n</ul>\n<ul>\n<li>Strong communication skills to collaborate with technical and non-technical stakeholders.</li>\n</ul>\n<ul>\n<li>Ability to work independently and collaboratively.</li>\n</ul>\n<p>Additional Preferred Skills &amp; Qualifications:</p>\n<ul>\n<li>Understanding of the financial markets, including experience with financial datasets, is strongly preferred.</li>\n</ul>\n<ul>\n<li>Experience with ML frameworks such as PyTorch, TensorFlow.</li>\n</ul>\n<ul>\n<li>Familiarity with MLOps practices and tools such as SageMaker, MLflow, or Airflow.</li>\n</ul>\n<ul>\n<li>Previous experience working in an Agile environment.</li>\n</ul>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package. The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1bd2d1b2-84f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"IT - Artificial Intelligence","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755954012324","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["Python","Machine Learning","NLP","GenAI","Agentic technologies","Data processing","Experimentation","Exploration","Cloud platforms (AWS, Azure, GCP)","Problem-solving skills","Communication skills"],"x-skills-preferred":["PyTorch","TensorFlow","MLOps practices and tools (SageMaker, MLflow, Airflow)","Agile environment"],"datePosted":"2026-04-18T22:14:27.951Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Machine Learning, NLP, GenAI, Agentic technologies, Data processing, Experimentation, Exploration, Cloud platforms (AWS, Azure, GCP), Problem-solving skills, Communication skills, PyTorch, TensorFlow, MLOps practices and tools (SageMaker, MLflow, Airflow), Agile environment","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_87749959-700"},"title":"Intern Data Engineering (all genders)","description":"<p>Join our Data Engineering team inside the Business Intelligence department, where you&#39;ll work with experienced engineers to build the data foundation that powers Holidu&#39;s growth.</p>\n<p>As an intern, you&#39;ll get hands-on experience with real problems and have the opportunity to make a meaningful impact. You&#39;ll work on building and supporting data pipelines, digging into data quality, getting hands-on with cloud infrastructure, and exploring AI-assisted development.</p>\n<p>Our team uses a range of technologies, including Redshift, Athena, DuckDB, Terraform, Docker, Jenkins, ELK, Grafana, Looker, OpsGenie, Kafka, Airbyte, and Fivetran. You&#39;ll have the chance to learn from experienced engineers and contribute to the development of our data systems.</p>\n<p>In this role, you&#39;ll be part of a team that genuinely loves what they do and is passionate about building a better data foundation for Holidu. You&#39;ll have the opportunity to take responsibility from day one and develop through regular feedback.</p>\n<p>We offer a fair salary, the chance to make a difference for hundreds of thousands of monthly users, and the opportunity to grow and develop through regular feedback. You&#39;ll also have access to a range of benefits, including a hybrid work policy, the chance to work from other local offices, and a corporate subscription to Urban Sports Club or a premium gym membership at a discounted rate.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_87749959-700","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2557398","x-work-arrangement":"hybrid","x-experience-level":"intern","x-job-type":"Internship","x-salary-range":null,"x-skills-required":["Python","SQL","Git","Airflow","dbt","Docker","Cloud platform (AWS, GCP, etc.)"],"x-skills-preferred":["LLM tools","AI-assisted coding"],"datePosted":"2026-04-18T22:13:52.778Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Git, Airflow, dbt, Docker, Cloud platform (AWS, GCP, etc.), LLM tools, AI-assisted coding"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ad717304-da7"},"title":"Intern Data Analytics (all genders)","description":"<p>You will be part of the Business Intelligence department, which consists of the Data Science, Data Analytics, and Data Engineering teams.</p>\n<p>This internship provides a great opportunity to gain hands-on experience into Data Analytics. You will work alongside a team of highly skilled and dedicated professionals who are committed to offering strong mentorship and guidance to help you start your career in the field of data.</p>\n<p>Duration: 6 months. Location: Munich, 2-3 office days per week.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git, Atlassian.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>As a Data Analytics Intern at Holidu, you’ll help our company make smarter, data-driven decisions, while being supported by a Senior Analyst.</p>\n<p>This role goes beyond building dashboards. We want curious, proactive people who want to become data advisors - not only delivering reports, but understanding the business context, which questions they answer and why they matter.</p>\n<ul>\n<li>Collect, analyse, and interpret large datasets to help solve real business challenges.</li>\n<li>Build dashboards and reports using tools like SQL, Python, and Looker.</li>\n<li>Collaborate closely with teams such as Product, Marketing, or Finance to help them extract actionable insights from data.</li>\n<li>Build and improve data pipelines using cutting-edge technologies.</li>\n<li>We are an AI-first team. Rather than manually executing repetitive tasks, you will use AI to work smarter and automate workflows.</li>\n<li>You’ll collaborate with our Data Scientists and get exposure to:</li>\n<li>Data preparation and exploratory data analysis.</li>\n<li>How ML-models are built, evaluated, and deployed in real-life.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>Currently enrolled in or recently completed a Bachelor’s or Master’s degree in a quantitative field (e.g., Business Analytics, Data Science, Economics, Statistics, Mathematics, Engineering or similar).</li>\n<li>Understanding of SQL and Python, proficiency in Excel/Google Sheets and a desire to learn visualization tools like Looker.</li>\n<li>Knowledge of Machine Learning and Statistical models is a plus.</li>\n<li>Strong analytical and problem-solving skills, and attention to detail.</li>\n<li>Curiosity to learn and a passion for solving data problems.</li>\n<li>Good communication and presentation skills.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Compensation: Get a fair salary.</li>\n<li>Impact: Make a difference for hundreds of thousands of monthly users.</li>\n<li>Growth: Take responsibility from day one and develop through regular feedback.</li>\n<li>Community: Engage with international, diverse, yet like-minded colleagues through regular events and 2 office days per week with your team.</li>\n<li>Flexibility: Benefit from our hybrid work policy and the chance to work from other local offices for up to 8 weeks a year.</li>\n<li>Fitness: Get a Urban Sports Club corporate subscription or a premium gym membership at a discounted rate.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ad717304-da7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2556233","x-work-arrangement":"hybrid","x-experience-level":"intern","x-job-type":"Internship","x-salary-range":null,"x-skills-required":["SQL","Python","Looker","Git","Atlassian","Airflow","DBT","AWS Stack","Redshift","Athena","Glue","S3"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:45.423Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Looker, Git, Atlassian, Airflow, DBT, AWS Stack, Redshift, Athena, Glue, S3"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_cc9213ff-135"},"title":"(Senior) Team Lead Marketing Analytics (all genders)","description":"<p>Within the Marketing Technology department, we are building a new Marketing Analytics team and are looking for a Team Lead to shape it from the ground up.</p>\n<p>You&#39;ll work closely with a wide range of Marketing stakeholders, ensuring they have the data, tools, and insights they need to drive sustainable growth. Moreover, you will also collaborate with data scientists and data engineers within the department to build best-in-class analytical solutions.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git, Jira, Confluence, Slack.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<ul>\n<li>You&#39;ll be leading data analysts and collaborating cross-functionally with data engineers and data scientists - fostering collaboration, learning, and analytical excellence.</li>\n<li>Engage with senior marketing leadership on strategic projects, providing insights that influence channel strategy and budget decisions, and ultimately our revenue growth.</li>\n<li>Translate marketing logic, for a diverse range of channels (e.g. Performance Marketing, SEO, CRM, affiliate) and use cases into analytical requirements and communicate complex findings clearly to both technical and commercial teams.</li>\n<li>Support and partner with Marketing Technology on tracking, event design, and data flows to ensure data quality and reliable reporting frameworks.</li>\n<li>Not shying away from hands-on work as an individual contributor (50% to start), while leading the team, diving deep into the details when needed.</li>\n<li>Shape the future of marketing analytics at Holidu by recruiting top talent, setting clear goals, and developing your team personally and professionally.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>5+ years multi-channel marketing analytics experience in a B2B or B2C organisation where marketing is a core performance driver, with extensive hands-on expertise in at least one of the following: attribution, cost and revenue allocation, or bidding.</li>\n<li>People management experience - this should not be your first leadership role.</li>\n<li>A collaborative mindset with clear experience communicating with executive stakeholders and senior decision makers.</li>\n<li>You are mission-driven, with a working backwards mentality (i.e. starting with customer needs) and clear experience managing and delivering complex projects with multiple stakeholders. Ability to translate business goals into analytical solutions and break down complex topics into actionable insights.</li>\n<li>Excellent analytical and technical skills. Concretely: strong in SQL, Python (or similar), data visualisation skills as well as developing technical frameworks to serve a clear business need.</li>\n<li>A strong personal or team focus on AI enablement: you actively use AI tools to enhance your coding, planning, and workflows, and can enable your team to do the same.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_cc9213ff-135","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2458940","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["AWS Stack","Airflow","DBT","Looker","SQL","Python","Git","Jira","Confluence","Slack"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:45.213Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Marketing","industry":"Technology","skills":"AWS Stack, Airflow, DBT, Looker, SQL, Python, Git, Jira, Confluence, Slack"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6690b2fa-cab"},"title":"(Senior) Team Lead Data Analytics (all genders)","description":"<p>At Holidu, data isn&#39;t just a support function, it&#39;s how we make decisions. The Analytics team builds the products and foundations that keep the whole organisation sharp, from day-to-day operations to long-term strategy.</p>\n<p>This role is on-site in Munich, with two office days per week.</p>\n<p>As a Senior Team Lead Data Analytics, you will lead one of Holidu&#39;s core analytics teams, a function at the intersection of data, strategy, and real business impact. The team has four direct reports and entails collaborating cross-functionally with data engineers and data scientists.</p>\n<p>Engage with senior leadership on strategic projects, providing insights that influence product strategy, internal operations, and revenue growth.</p>\n<p>You and your team will support a range of stakeholders across the company (e.g. Customer Support, Host Experience, Sales and Account Management).</p>\n<p>As a member of the BI leadership team, you will help shape the department strategy and the future of AI-powered data products.</p>\n<p>Understand problems and identify opportunities across a diverse range of stakeholder use cases, translating them into analytical requirements and communicating complex findings clearly to both technical and commercial audiences.</p>\n<p>Lead from the front: this role carries meaningful individual contributor responsibility. You&#39;ll be expected to do real analytical work, diving deep into the data, building solutions, and setting the bar for quality in your team.</p>\n<p>Shape the future of analytics at Holidu by recruiting top talent, setting clear goals, and developing your team personally and professionally.</p>\n<p>The ideal candidate will have 5+ years of data analytics experience, people management experience, a collaborative mindset, a mission-driven mentality, excellent analytical and technical skills, and a genuine commitment to AI enablement.</p>\n<p>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</p>\n<p>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</p>\n<p>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</p>\n<p>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</p>\n<p>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</p>\n<p>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6690b2fa-cab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2598226","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Database: AWS Stack (Redshift, Athena, Glue, S3)","Data Pipelines: Airflow, dbt","Data Visualisation: Looker","Data Analytics: SQL, Python","Collaboration: Git, Jira, Confluence, Slack"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:28.264Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Technology","industry":"Travel Technology","skills":"Database: AWS Stack (Redshift, Athena, Glue, S3), Data Pipelines: Airflow, dbt, Data Visualisation: Looker, Data Analytics: SQL, Python, Collaboration: Git, Jira, Confluence, Slack"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_25fd58ed-3c0"},"title":"(Senior) Data Scientist (all genders)","description":"<p>You will be part of the Business Intelligence department, which consists of the Data Science, Data Analytics, and Data Engineering teams.</p>\n<p>As a Senior Data Scientist, you will work on various topics such as rankings, recommendations, user segmentation, user lifetime value, business forecasts, etc. You will have access to our huge dataset and work in collaboration with stakeholders from various departments.</p>\n<p>Your objective is to build the best internal and external products for our customers. Holidu highly values a diverse and open environment with people from all over the world.</p>\n<p>This role is based in Munich with a hybrid setup.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Flexible data science environment (Python, Sagemaker)</li>\n<li>Database: AWS Stack (Redshift, Athena, Glue, S3).</li>\n<li>Data Pipelines: Airflow, DBT.</li>\n<li>Data Visualization: Looker.</li>\n<li>Data Analytics: SQL, Python.</li>\n<li>Collaboration: Git.</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>You will play a pivotal role in the Business Intelligence team alongside data scientists, analysts, and engineers. Together, you will lead the development and enhancement of our company-wide machine learning strategy.</p>\n<ul>\n<li>Collaborate across various business departments to identify opportunities and solve critical business challenges using data science solutions.</li>\n<li>Build and optimize predictive models such as booking cancellation forecasts, churn predictions, pricing optimization, revenue forecasting and marketing channel allocation.</li>\n<li>Take models from conception to production, continuously monitor their performance, and iterate to enhance accuracy and efficiency.</li>\n<li>Interface with diverse business stakeholders, ensuring alignment between data science initiatives and company goals.</li>\n<li>Demonstrate leadership in data science projects, leveraging your expertise to drive measurable business impact.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>3+ years of experience as a Data Scientist, with a proven track record of applying data science methodologies to solve complex business problems.</li>\n<li>A degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field.</li>\n<li>Expertise in statistics, predictive analytics, machine learning techniques, and proficiency in tools like Python and SQL.</li>\n<li>Experience with Airflow and dbt is a plus.</li>\n<li>Strong understanding of business operations and experience collaborating with diverse stakeholders.</li>\n<li>Enthusiasm for data science and a drive to deliver world-class products that make a difference.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts.</li>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback.</li>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other.</li>\n<li>Technology: Work in a modern tech environment.</li>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations.</li>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_25fd58ed-3c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2555141","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","Sagemaker","AWS Stack","Airflow","DBT","Looker","SQL","Git"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:13:07.588Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Sagemaker, AWS Stack, Airflow, DBT, Looker, SQL, Git"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8610ea3d-93b"},"title":"Cloud Platform Engineer","description":"<p>The Business Development/Management Technology team at FIC &amp; Risk Technology is building and operating platforms that support recruiting, hiring, and onboarding of investment professionals. We are currently integrating multiple legacy and new systems into a unified, cloud-native platform to standardize processes, workflows, and data models across the organisation.</p>\n<p>This integration will enable seamless collaboration between teams and provide reliable, scalable data for analytics and reporting. We are looking for a Cloud Platform Engineer to design, build, and operate our AWS-based infrastructure and data platforms, using modern DevOps practices, infrastructure as code, and secure, well-engineered services in Python and C#.</p>\n<p>The successful candidate will collaborate with global technology and business teams to design cloud-native solutions that support business development and onboarding workflows. They will partner with global stakeholders to understand requirements and translate them into secure, scalable AWS architectures and platform capabilities.</p>\n<p>Key responsibilities include leading the end-to-end delivery of cloud and platform features, including design, implementation (Python/C#), infrastructure as code, testing, and deployment using DevOps practices.</p>\n<p>We are looking for a highly skilled engineer with 6+ years of experience in software or platform engineering, with significant time spent building and operating solutions in cloud environments (AWS preferred).</p>\n<p>The ideal candidate will have strong hands-on programming experience in Python and C#, with solid understanding of object-oriented design, design patterns, service-oriented / microservices architectures, concurrency, and SOLID principles.</p>\n<p>They will also have proven experience designing and operating AWS-based platforms (e.g., EC2, ECS/EKS, Lambda, S3, RDS, IAM) using infrastructure as code (Terraform, CloudFormation, or CDK).</p>\n<p>In addition, the successful candidate will have practical experience implementing DevOps practices and CI/CD pipelines (e.g., Jenkins, GitHub Actions, Azure DevOps), including automated testing, security scanning, and deployment.</p>\n<p>Experience supporting data science and analytics platforms, including orchestration tools such as Airflow, distributed processing engines such as Spark, and cloud-native data pipelines is also required.</p>\n<p>Good understanding of SQL and core database concepts; familiarity with AWS analytics services (e.g., Glue, EMR, Redshift, Athena) is a plus.</p>\n<p>Awareness of cloud security best practices, including IAM, network security, data encryption, and secure configuration management is also necessary.</p>\n<p>Strong problem-solving and analytical skills; demonstrated ability to take ownership, deliver in a fast-paced environment, and collaborate effectively with global teams is essential.</p>\n<p>Excellent communication skills, with ability to work closely with both technical and non-technical stakeholders is also required.</p>\n<p>Experience estimating, monitoring, and optimizing AWS infrastructure costs, including use of tools such as AWS Cost Explorer, AWS Budgets, and cost-allocation tagging strategies is desirable.</p>\n<p>Experience designing and operating workloads across multiple cloud environments and on-premises, using centralized policies, governance, and controls to support business-aligned teams is also beneficial.</p>\n<p>Working knowledge of networking across on-premises and cloud environments, including VPC design, subnets, routing, VPNs/Direct Connect, load balancing, DNS, and network security controls is necessary.</p>\n<p>Nice to have experience with additional big data tools or platforms (e.g., Kafka, Databricks, Snowflake, Flink).</p>\n<p>Familiarity with Capital Markets concepts and operating models is also beneficial.</p>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p>Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p>When finalising an offer, we take into consideration an individual&#39;s experience level and the qualifications they bring to the role to formulate a competitive total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8610ea3d-93b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"FIC & Risk Technology","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955139979","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["AWS","Python","C#","DevOps","Infrastructure as Code","Cloud Security","SQL","Database Concepts","Networking"],"x-skills-preferred":["Airflow","Spark","Kafka","Databricks","Snowflake","Flink","Capital Markets"],"datePosted":"2026-04-18T22:12:50.548Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"AWS, Python, C#, DevOps, Infrastructure as Code, Cloud Security, SQL, Database Concepts, Networking, Airflow, Spark, Kafka, Databricks, Snowflake, Flink, Capital Markets","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9ca997fb-218"},"title":"Quantitative Developer","description":"<p>We are building a world-class systematic data platform that will power the next generation of our systematic portfolio engines.</p>\n<p>The systematic data group is looking for a Quantitative Developer to join our growing team. The team consists of content specialists, data scientists, engineers, and quant developers who are responsible for discovering, maintaining, and analysing sources of alpha for our portfolio managers.</p>\n<p>The role builds on individual&#39;s knowledge and skills in four key areas of quantitative investing: data, statistics, technology, and financial markets.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Use finance knowledge and statistical knowledge to analyse potential alpha sources and present findings to portfolio managers and quantitative analysts.</li>\n<li>Build quant tools to help portfolio managers research, evaluate, combine alphas, and understand risks.</li>\n<li>Design and maintain tools to evaluate and monitor data quality and integrity for a wide variety of data sources.</li>\n<li>Engage with vendors, brokers, and perform analytics to understand characteristics of datasets.</li>\n<li>Interact with portfolio managers and quantitative analysts to understand their use cases and recommend datasets to help maximise their profitability.</li>\n</ul>\n<p>Skills Required:</p>\n<ul>\n<li>3+ years of work experience as a financial engineer, data scientist, or quant developer.</li>\n<li>Strong knowledge of Python and/or C++, Java, C#.</li>\n<li>Familiarity with data pipeline engineering, ETL for large datasets, and scheduling tools like Airflow.</li>\n<li>Strong SQL and database experience including PL-SQL or T-SQL.</li>\n<li>Understanding of typical software development lifecycle and familiarity with: Linux, GitHub, CI/CD.</li>\n<li>Ph.D. or Masters in computer science, mathematics, statistics, or other field requiring quantitative analysis.</li>\n</ul>\n<p>Beneficial Skills and Experience:</p>\n<ul>\n<li>Understanding of risk models and performance attribution.</li>\n<li>Experience with financial markets such as equities and futures.</li>\n<li>Knowledge of statistical techniques and their usage.</li>\n</ul>\n<p>The estimated base salary range for this position is $165,000 to $250,000, which is specific to New York and may change in the future. Millennium pays a total compensation package which includes a base salary, discretionary performance bonus, and a comprehensive benefits package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9ca997fb-218","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Equity IT","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755952876477","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$165,000 to $250,000","x-skills-required":["Python","C++","Java","C#","data pipeline engineering","ETL","Airflow","SQL","database","Linux","GitHub","CI/CD","Ph.D.","Masters"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:44.538Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, C++, Java, C#, data pipeline engineering, ETL, Airflow, SQL, database, Linux, GitHub, CI/CD, Ph.D., Masters","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_21f5f6c3-734"},"title":"Data Engineer","description":"<p>About the Role We are at a pivotal scaling point where our data ambitions have outpaced our current setup, and we need a Data Engineer to architect the professional-grade foundations of our platform.</p>\n<p>This role exists to bridge the gap between &quot;getting data&quot; and &quot;engineering data,&quot; moving us from manual syncs to a fully automated ecosystem. By building custom pipelines and implementing a robust orchestration layer, you will directly enable our Operations teams and leadership to transition from basic reporting to sophisticated, AI-ready data products.</p>\n<p>Your primary focus will be on Infrastructure-as-Code, orchestration, and building a resilient &quot;plumbing&quot; system that serves as the backbone for our entire Product and GTM strategy.</p>\n<p>Your 12-Month Journey During the first 3 months: you will learn about our existing stack (GCP, BigQuery, Airbyte, dbt) and understand the current pain points in our data flow. You will identify and execute &quot;low-hanging fruit&quot; improvements to our product usage analytics, providing immediate value to the Product and GTM teams. You’ll begin designing the blueprint for our custom data pipelines and the migration strategy for moving our infrastructure into Terraform.</p>\n<p>Within 6 months: You will have deployed our new orchestration layer (e.g., Airflow or Dagster) and successfully transitioned our first set of custom pipelines to production. Collaborating with the Analytics Engineer, you will enable a unified view of our customer journey by successfully merging product usage data with CRM and billing data. At this point, a significant portion of our data infrastructure will be defined as code, reducing manual overhead and increasing deployment reliability.</p>\n<p>After 1 year: you will take full strategic ownership of the data platform and its long-term architecture. You will act as the go-to technical expert for the leadership team, advising on the scalability of new data-driven features. You will lay the groundwork for AI and Machine Learning initiatives by ensuring our data warehouse has the right quality controls, governance, and low-latency access patterns in place.</p>\n<p>What You’ll Be Doing Architect Scalable Infrastructure-as-Code: Take our existing foundations to the next level by migrating all GCP and BigQuery resources into Terraform. You will establish automated CI/CD patterns to ensure our entire data environment is reproducible, version-controlled, and enterprise-ready.</p>\n<p>Deploy State-of-the-Art Pipelines: Design, deploy, and operate high-quality production ELT pipelines. You will implement a modern orchestration layer (e.g., Airflow or Dagster) to build custom Python-based integrations while maintaining and optimizing our existing syncs.</p>\n<p>Champion Data Quality &amp; Performance: Act as the guardian of our data platform. You will implement rigorous testing and monitoring protocols to ensure data is accurate and timely. You will proactively identify BigQuery bottlenecks, optimizing query performance and resource utilization.</p>\n<p>Technical Roadmap &amp; Ownership: scope and architect end-to-end data flows from production source to warehouse. Manage your own technical backlog, prioritizing infrastructure stability over technical debt. You will ensure platform security and SOC2 compliance through PII masking, data contracts, and robust access controls.</p>\n<p>Collaboration: You will work in a tight loop with the Analytics Engineer to turn raw data into actionable products. You will partner daily with DataOps and RevOps to understand business requirements, with occasional strategic syncs with DevOps and R&amp;D to align on production schema changes and global infrastructure standards.</p>\n<p>What You Bring Solid experience in Data Engineering, with a track record of building and evolving data ingestion infrastructure in cloud environments. The Modern Data Stack: Familiarity with dbt and Airbyte/Fivetran. You understand how these tools fit into a broader ecosystem. Expertise in BigQuery (partitioning, clustering, IAM) and the broader GCP ecosystem; Infrastructure-as-Code (Terraform). Hands-on experience with Airflow, Dagster, or similar orchestration tools. You know how to design DAGs that are resilient and easy to debug. DevOps practices in the data context: familiarity with CI/CD best practices as they apply to data (data testing, automated deployments). Programming: Expert-level Python and advanced SQL. You are comfortable writing clean, testable, and modular code. Comfortable in a fast-paced environment Project management skills: capable of managing stakeholders, explaining complicated technical trade-offs to non-technical users, and taking care of own project scoping and backlog management. Fluency in English, both written and spoken, at a minimum C1 level</p>\n<p>What We Offer Flexibility to work from home in the Netherlands and from our beautiful canal-side office in Amsterdam A chance to be part of and shape one of the most ambitious scale-ups in Europe Work in a diverse and multicultural team €1,500 annual training budget plus internal training Pension plan, travel reimbursement, and wellness perks 28 paid holiday days + 2 additional days to relax in 2026 Work from anywhere for 4 weeks/year An inclusive and international work environment with a whole lot of fun thrown in! Apple MacBook and tools €200 Home Office budget</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_21f5f6c3-734","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Tellent","sameAs":"https://careers.tellent.com","logo":"https://logos.yubhub.co/careers.tellent.com.png"},"x-apply-url":"https://careers.tellent.com/o/data-engineer","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"EUR 70000–90000 / year","x-skills-required":["Data Engineering","Cloud environments","dbt","Airbyte/Fivetran","BigQuery","GCP ecosystem","Infrastructure-as-Code","Terraform","Airflow","Dagster","Python","SQL","CI/CD best practices","DevOps practices"],"x-skills-preferred":[],"datePosted":"2026-04-18T22:12:06.548Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Amsterdam"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Data Engineering, Cloud environments, dbt, Airbyte/Fivetran, BigQuery, GCP ecosystem, Infrastructure-as-Code, Terraform, Airflow, Dagster, Python, SQL, CI/CD best practices, DevOps practices","baseSalary":{"@type":"MonetaryAmount","currency":"EUR","value":{"@type":"QuantitativeValue","minValue":70000,"maxValue":90000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6bfc6b4-74f"},"title":"Senior Data Scientist - Marketing (all genders)","description":"<p>Join our Business Intelligence Department, a multidisciplinary group of Data Scientists, Analysts, and Data Engineers. Together, we build machine learning and analytics products that directly influence GMV, conversion, and retention.</p>\n<p>Within the department, we’re building a new Marketing Analytics team and are looking for a Senior Data Scientist to drive its data science initiatives. In this role, you’ll work closely with Analysts, Engineers, and Marketing stakeholders to develop and productionize advanced machine learning, statistical, and predictive models that improve marketing performance and drive measurable company growth.</p>\n<p>As a Senior Data Scientist – Marketing, you’ll take strong ownership of data science initiatives that directly shape our marketing strategy and growth. You will:</p>\n<p>Partner closely with Marketing, Marketing Analytics, and Marketing Technology to identify opportunities and translate business questions into scalable data science solutions.</p>\n<p>Lead the development of high-impact machine learning and statistical models for marketing use cases such as channel allocation, ad bidding, churn prediction, lifetime value, revenue attribution, and business metrics forecasting.</p>\n<p>Work end-to-end - from translating business questions into hypotheses to researching, building, validating, and deploying models.</p>\n<p>Run experiments and iterate in production: design A/B tests, monitor model performance, and continuously improve based on measured impact.</p>\n<p>Advance our MLOps practices with CI/CD pipelines, retraining workflows, lineage tracking, and documentation.</p>\n<p>Help define the team&#39;s roadmap and ways of working as a founding member of Marketing Analytics - your input will help shape this function.</p>\n<p>Act as a senior role model in the team, sharing best practices and helping raise the bar for data science at Holidu.</p>\n<p>We&#39;re looking for someone with 5+ years of experience as a Data Scientist, with clear ownership of projects that delivered measurable business impact. You should have a degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field, and strong expertise in machine learning, statistics, and predictive analytics, with hands-on experience using Python and SQL.</p>\n<p>Experience with marketing data science use cases such as attribution modeling, customer lifetime value prediction, churn modeling, or bid optimization is also required. You should have a solid understanding of marketing concepts across channels (e.g. Performance Marketing, SEO, CRM, Affiliate) and how data science can improve them.</p>\n<p>Additionally, you should have experience working with modern data stacks, ideally including AWS (Redshift, Athena, S3), Airflow, dbt, and Git. A collaborative mindset paired with great communication skills is essential, as you&#39;ll need to work with diverse stakeholders and explain complex topics in a simple way.</p>\n<p>AI proficiency is also a plus, as you&#39;ll be comfortable using AI to enhance coding, planning, and monitoring, and successfully integrating AI tools (such as Claude code, Codex, Copilot, etc.) into your workflow and teaching others to use them efficiently.</p>\n<p>If you&#39;re excited about the opportunity to shape the future of travel with products used by millions of guests and thousands of hosts, apply now!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6bfc6b4-74f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2510157","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Machine Learning","Statistics","Predictive Analytics","Python","SQL","Marketing Data Science","Attribution Modeling","Customer Lifetime Value Prediction","Churn Modeling","Bid Optimization"],"x-skills-preferred":["AI","CI/CD Pipelines","Retraining Workflows","Lineage Tracking","Documentation","Airflow","dbt","Git"],"datePosted":"2026-04-18T22:10:24.739Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, Statistics, Predictive Analytics, Python, SQL, Marketing Data Science, Attribution Modeling, Customer Lifetime Value Prediction, Churn Modeling, Bid Optimization, AI, CI/CD Pipelines, Retraining Workflows, Lineage Tracking, Documentation, Airflow, dbt, Git"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_80d15de9-aa7"},"title":"Senior Data Scientist - Rankings & Recommendations (all genders)","description":"<p>Join our Business Intelligence Department, a multidisciplinary group of Data Scientists, Analysts, and Data Engineers.</p>\n<p>You will join a cross-functional Product team, Search Intelligence, which is responsible for optimizing ranking and recommendations for users visiting our website.</p>\n<p>You&#39;ll be part of the broader Data Science team, which operates across cross-functional domain teams - giving you access to shared knowledge, best practices, and collaboration opportunities beyond your domain.</p>\n<p>You’ll collaborate daily with Data Engineers, Analysts, Product Managers, and Back-end Engineers.</p>\n<p>You’ll report to the Team Lead, Data Science.</p>\n<p>Together, we turn data into actionable insights and innovative technology that powers how millions of guests find and book their perfect holiday home.</p>\n<p><strong>Our Tech Stack</strong></p>\n<ul>\n<li>Python • Airflow • dbt • AWS (SageMaker, Redshift, Athena) • MLflow</li>\n</ul>\n<p>The Ranking challenge at Holidu</p>\n<p>Holidu lists over 4 million vacation rental properties. Our ranking and personalization systems determine which of them our 70+ million annual users see, directly impacting search conversion and business results.</p>\n<p>What&#39;s live today:</p>\n<ul>\n<li>Multi-stage ranking pipeline: Reinforcement-learning-based cold ranking, contextual re-ranking, and personalized recommendations.</li>\n</ul>\n<ul>\n<li>Cold-start models for new properties with limited behavioral data.</li>\n</ul>\n<ul>\n<li>Personalized recommendations based on user browsing patterns.</li>\n</ul>\n<p>Some of the hard problems we&#39;re solving:</p>\n<ul>\n<li>Multi-objective optimization: Balancing user relevance, conversion probability, and business value.</li>\n</ul>\n<ul>\n<li>Personalization without history: Most users are anonymous or first-time visitors.</li>\n</ul>\n<ul>\n<li>Cold-start: A significant share of our inventory is new each quarter. How do we surface quality properties before we have behavioral data?</li>\n</ul>\n<p><strong>Your role in this journey</strong></p>\n<p>You&#39;ll shape the ranking and recommendation systems that millions of guests rely on to find their holiday home. With access to extensive datasets and modern ML infrastructure, you&#39;ll work end-to-end - from identifying opportunities and prototyping new approaches to shipping models to production and measuring their impact.</p>\n<ul>\n<li>Develop high-impact models and improvements for our ranking, recommendation, and personalization systems - with the freedom to explore new, creative approaches.</li>\n</ul>\n<ul>\n<li>Take models from conception to production, continuously monitor their performance, and iterate to enhance accuracy and efficiency.</li>\n</ul>\n<ul>\n<li>Design and run A/B tests as a core part of ranking development; success is measured by successful experiments per quarter and time-to-decision.</li>\n</ul>\n<ul>\n<li>Collaborate closely with Product Managers and Software Engineers to identify, prioritize, and ship ranking improvements.</li>\n</ul>\n<ul>\n<li>Ensure model reliability in production, measured by online/offline agreement, model and data drift KPIs, latency and uptime SLAs, and automated monitoring coverage.</li>\n</ul>\n<ul>\n<li>Advance our MLOps practices with CI/CD pipelines, retraining workflows, lineage tracking, and documentation.</li>\n</ul>\n<ul>\n<li>Demonstrate leadership in data science projects by driving technical direction, scoping initiatives, and guiding the team&#39;s prioritization and project execution.</li>\n</ul>\n<p><strong>Your backpack is filled with</strong></p>\n<ul>\n<li>5+ years of experience as a Data Scientist, with a proven track record of applying ML models to solve real business problems.</li>\n</ul>\n<ul>\n<li>Experience working on ranking models or recommender systems is a strong advantage.</li>\n</ul>\n<ul>\n<li>A degree in Machine Learning, Computer Science, Mathematics, Physics, or a related field.</li>\n</ul>\n<ul>\n<li>Strong foundations in statistics, predictive modeling, and machine learning techniques, with hands-on experience using Python and SQL.</li>\n</ul>\n<ul>\n<li>Experience with Airflow and dbt is a plus.</li>\n</ul>\n<ul>\n<li>Solid understanding of business operations and the ability to translate data insights into clear, actionable outcomes.</li>\n</ul>\n<ul>\n<li>A collaborative mindset and enthusiasm for using data to build world-class products that make a real impact.</li>\n</ul>\n<ul>\n<li>AI Proficiency: You are comfortable using AI to enhance coding, planning, and monitoring. This includes successfully integrating AI tools (such as Claude code, Codex, Copilot, etc.) into your workflow and teaching others to use them efficiently.</li>\n</ul>\n<p><strong>Our adventure includes</strong></p>\n<ul>\n<li>Impact: Shape the future of travel with products used by millions of guests and thousands of hosts. At Holidu ideas become products, data drives decisions, and iteration fuels fast learning. Your work matters - and you’ll see the impact.</li>\n</ul>\n<ul>\n<li>Learning: Grow professionally in a culture that thrives on curiosity and feedback. You’ll learn from outstanding colleagues, collaborate across disciplines, and benefit from mentorship, and personal learning budgets - with a strong focus on AI.</li>\n</ul>\n<ul>\n<li>Great People: Join a team of smart, motivated and international colleagues who challenge and support each other. We celebrate wins and keep our culture fun, ambitious and human. Our customers are guests and hosts - people we can all relate to - making work meaningful and energizing.</li>\n</ul>\n<ul>\n<li>Technology: Work in a modern tech environment. You’ll experience the pace of a scale-up combined with the stability of a proven business model, enabling you to build, test, and improve continuously.</li>\n</ul>\n<ul>\n<li>Flexibility: Work a hybrid setup with 50% in-office time for collaboration, and spend up to 8 weeks a year from other inspiring locations. You’ll stay connected through regular events and meet-ups across our almost 30 offices.</li>\n</ul>\n<ul>\n<li>Perks on Top: Of course, we also offer travel benefits, gym discounts, and other perks to keep you energized - but what truly sets us apart is the chance to grow in a dynamic industry, alongside amazing people, while having fun along the way.</li>\n</ul>\n<p>Need a sneak peek? Check out the adventure that awaits you on Instagram @lifeatholidu and dive straight into the world of Tech at Holidu for more insights!</p>\n<p><strong>Want to travel with us?</strong></p>\n<p>Apply online on our careers page! Your first travel contact will be Lucia from HR.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_80d15de9-aa7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Holidu Hosts GmbH","sameAs":"https://holidu.jobs.personio.com","logo":"https://logos.yubhub.co/holidu.jobs.personio.com.png"},"x-apply-url":"https://holidu.jobs.personio.com/job/2413808","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full-time","x-salary-range":null,"x-skills-required":["Python","Airflow","dbt","AWS","MLflow","Machine Learning","Statistics","Predictive Modeling","SQL"],"x-skills-preferred":["AI","Data Science","Ranking Models","Recommender Systems","Collaboration","Communication"],"datePosted":"2026-04-18T22:09:15.403Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Munich, Germany"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Airflow, dbt, AWS, MLflow, Machine Learning, Statistics, Predictive Modeling, SQL, AI, Data Science, Ranking Models, Recommender Systems, Collaboration, Communication"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6365e7d7-511"},"title":"Senior Forward Deployed Data Scientist/Engineer","description":"<p>We&#39;re hiring a Senior Forward Deployed Data Scientist / Engineer to work directly with customers on ambiguous, high-impact problems at the intersection of data science, product development, and AI deployment.</p>\n<p>This is not a traditional analytics role. On this team, data scientists do the core statistical and modeling work, but they also build real tools and products: evaluation explorers, operator workflows, decision-support systems, experimentation surfaces, and customer-specific AI/data applications that get used in production.</p>\n<p>The right candidate is strong in first-principles problem solving, rigorous measurement, and technical execution. They know how to define metrics, design experiments, diagnose failures, and build systems that people actually use. They are also comfortable using modern AI-assisted development tools to prototype and iterate quickly without sacrificing reliability, observability, or judgment. Python and SQL matter in this role, but as execution fluency in service of building better products and making better decisions.</p>\n<p>Responsibilities: Partner directly with enterprise customers to understand workflows, operational pain points, constraints, and success criteria Turn ambiguous business and product problems into measurable solutions with clear metrics, technical designs, and deployment plans Design and build internal and customer-facing data products, including evaluation tools, workflow applications, decision-support systems, and thin product layers on top of data/ML systems Build end-to-end solutions across data ingestion, transformation, experimentation, statistical modeling, deployment, monitoring, and iteration Design evaluation frameworks, benchmarks, and feedback loops for ML/LLM systems, human-in-the-loop workflows, and model-assisted operations Apply rigorous statistical thinking to experimentation, causal inference, metric design, forecasting, segmentation, diagnostics, and performance measurement Use AI-assisted development workflows to accelerate prototyping and product iteration, while maintaining strong engineering discipline Diagnose failure modes across data quality, model behavior, retrieval, workflow design, and user experience, and drive fixes into production Act as the voice of the customer to Product, Engineering, and Data Science, using field learnings to shape roadmap and platform capabilities</p>\n<p>Requirements: 5+ years of experience in data science, machine learning, quantitative engineering, or another highly analytical technical role Proven track record of shipping data, ML, or AI systems that delivered measurable business or product impact Exceptional ability to structure ambiguous problems, define the right success metrics, and translate them into executable technical plans Strong foundation in statistics, experimentation, causal reasoning, and measurement Experience building tools or products, not just analyses , for example internal workflow tools, evaluation systems, operator-facing products, experimentation platforms, or customer-specific applications Hands-on fluency in Python, SQL, and modern data/AI tooling; able to inspect data, prototype quickly, debug deeply, and productionize solutions that work Comfort using AI-assisted coding and development workflows to move from idea to usable product quickly Strong communication and stakeholder management skills; able to work effectively with customers, engineers, product teams, and executives High ownership and bias toward shipping in fast-moving environments with incomplete information</p>\n<p>Preferred qualifications: Experience in a forward deployed, solutions, consulting, or other client-facing technical role Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign</p>\n<p>What success looks like: Success in this role means taking a messy, high-stakes customer problem and turning it into a deployed system that is actually used. Sometimes that system is a model. Sometimes it is an evaluation framework. Sometimes it is an operator-facing tool or a lightweight data product that changes how decisions get made. In all cases, success is defined by measurable impact, rigorous evaluation, and reliable execution.</p>\n<p>Compensation packages at Scale for eligible roles include base salary, equity, and benefits. The range displayed on each job posting reflects the minimum and maximum target for new hire salaries for the position, determined by work location and additional factors, including job-related skills, experience, interview performance, and relevant education or training. Scale employees in eligible roles are also granted equity based compensation, subject to Board of Director approval. Your recruiter can share more about the specific salary range for your preferred location during the hiring process, and confirm whether the hired role will be eligible for equity grant. You’ll also receive benefits including, but not limited to: Comprehensive health, dental and vision coverage, retirement benefits, a learning and development stipend, and generous PTO. Additionally, this role may be eligible for additional benefits such as a commuter stipend.</p>\n<p>Salary Range: $167,200-$209,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6365e7d7-511","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Scale AI","sameAs":"https://scale.com/","logo":"https://logos.yubhub.co/scale.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/scaleai/jobs/4636227005","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$167,200-$209,000 USD","x-skills-required":["Python","SQL","Modern data/AI tooling","Statistics","Experimentation","Causal reasoning","Measurement","Data science","Machine learning","Quantitative engineering"],"x-skills-preferred":["Experience in a forward deployed, solutions, consulting, or other client-facing technical role","Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products","Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow","Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery","Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems","Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling","Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign"],"datePosted":"2026-04-18T15:59:44.618Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; New York, NY"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Modern data/AI tooling, Statistics, Experimentation, Causal reasoning, Measurement, Data science, Machine learning, Quantitative engineering, Experience in a forward deployed, solutions, consulting, or other client-facing technical role, Experience designing evaluation frameworks for LLMs, retrieval systems, agentic workflows, or other AI-enabled products, Experience with large-scale data processing and distributed systems such as Spark, Ray, or Airflow, Experience with cloud infrastructure and modern data platforms such as AWS, GCP, Snowflake, or BigQuery, Experience building lightweight applications, APIs, internal tools, or workflow software on top of data/ML systems, Familiarity with marketplace experimentation, causal inference, forecasting, optimization, or advanced statistical modeling, Strong product instinct and the judgment to know when the right answer is a model, an experiment, a tool, or a workflow redesign","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":167200,"maxValue":209000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b68ff4cc-e74"},"title":"Data Engineer, Safeguards","description":"<p><strong>About the role</strong></p>\n<p>Anthropic is looking for a Data Engineer to join the Safeguards team and build the data foundations that keep our AI systems safe. The Safeguards team works to monitor models, prevent misuse, and ensure user well-being.</p>\n<p>You&#39;ll design and build the data pipelines, warehousing solutions, and analytical tooling that power our safety and trust efforts at scale. You&#39;ll work closely with engineers, data scientists, and policy teams to ensure the Safeguards organization has the data it needs to detect abuse patterns, measure the effectiveness of safety interventions, and make informed decisions about model behavior and enforcement.</p>\n<p>This is a high-impact role where your work will directly support Anthropic&#39;s mission to develop AI that is safe and beneficial.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, build, and maintain scalable data pipelines that support safety monitoring, abuse detection, and enforcement workflows</li>\n<li>Develop and optimize data models and warehousing solutions to enable efficient analysis of large-scale usage and safety data</li>\n<li>Build and maintain dashboards and reporting infrastructure that give Safeguards teams visibility into model behavior, misuse patterns, and enforcement outcomes</li>\n<li>Collaborate with engineers to integrate data from multiple sources , including model outputs, user reports, and automated classifiers , into a unified analytical layer</li>\n<li>Implement data quality frameworks, monitoring, and alerting to ensure the reliability of safety-critical data</li>\n<li>Partner with research teams to surface data insights that inform model improvements and safety interventions</li>\n<li>Develop self-service data tooling that enables stakeholders to explore safety data and generate reports independently</li>\n<li>Contribute to data governance practices, including access controls, retention policies, and privacy-compliant data handling</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have 3+ years of experience in data engineering, analytics engineering, or a related role</li>\n<li>Are proficient in SQL and Python, with experience building and maintaining ETL/ELT pipelines</li>\n<li>Have hands-on experience with modern data stack tools such as dbt, Airflow, Spark, or similar orchestration and transformation frameworks</li>\n<li>Have worked with cloud data platforms (BigQuery, Redshift, Snowflake, or similar)</li>\n<li>Are comfortable building dashboards and data visualizations using tools like Looker, Tableau, or Metabase</li>\n<li>Communicate clearly and can translate complex data concepts for both technical and non-technical audiences</li>\n<li>Are results-oriented, flexible, and willing to pick up slack even when it falls outside your job description</li>\n<li>Care about the societal impacts of AI and are motivated by safety work</li>\n</ul>\n<p><strong>Strong candidates may have:</strong></p>\n<ul>\n<li>Experience with trust &amp; safety, integrity, fraud, or abuse detection data systems</li>\n<li>Experience with large-scale event streaming systems (Kafka, Pub/Sub, Kinesis)</li>\n<li>Built data infrastructure that supports ML model monitoring or evaluation</li>\n<li>A background in statistical analysis, or experience collaborating closely with data scientists</li>\n<li>Developed internal tooling or self-service analytics platforms</li>\n</ul>\n<p><strong>Strong candidates need not have:</strong></p>\n<ul>\n<li>A formal degree in Computer Science or a related field , we value practical experience and demonstrated ability over credentials</li>\n<li>Prior experience in AI or machine learning , you&#39;ll learn the domain-specific context on the job</li>\n<li>Previous experience at an AI safety or research organization</li>\n<li>Deep expertise across every tool listed above , familiarity with a subset and a willingness to learn is enough</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p>Minimum education: Bachelor’s degree or an equivalent combination of education, training, and/or experience Required field of study: A field relevant to the role as demonstrated through coursework, training, or professional experience Minimum years of experience: Years of experience required will correlate with the internal job level requirements for the position Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices. Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact , advancing our long-term goals of steerable, trustworthy AI , rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills. The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preferences.</p>\n<p><strong>Come work with us!</strong></p>\n<p>Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a lovely office space in which to collaborate with colleagues.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b68ff4cc-e74","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5156057008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"£170,000-£220,000 GBP","x-skills-required":["SQL","Python","ETL/ELT pipelines","dbt","Airflow","Spark","cloud data platforms","BigQuery","Redshift","Snowflake","Looker","Tableau","Metabase"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:59:33.960Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, UK"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, ETL/ELT pipelines, dbt, Airflow, Spark, cloud data platforms, BigQuery, Redshift, Snowflake, Looker, Tableau, Metabase","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":170000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_467be5c4-940"},"title":"Machine Learning Engineer","description":"<p>We&#39;re looking for a Machine Learning Engineer to join our Ads Engineering team. As a Machine Learning Engineer at Reddit, you will design and build production ML systems that power core experiences across the platform, including personalized recommendations, search, and ranking systems, intelligent advertising systems, and large-scale machine learning pipelines.</p>\n<p>Our team works on high-impact systems that operate at internet scale and directly influence user experience, advertiser value, and business outcomes. You&#39;ll work on complex, real-world ML problems at massive scale, and contribute to technical strategy, architecture, and long-term ML roadmap.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and deploy production-grade machine learning models and systems at scale</li>\n<li>Own the full ML lifecycle: from problem definition and feature engineering to training, evaluation, deployment, and monitoring</li>\n<li>Build scalable data and model pipelines with strong reliability, observability, and automated retraining</li>\n<li>Work with large-scale datasets to improve ranking, recommendations, search relevance, prediction, content/user understanding, and optimization systems</li>\n<li>Partner cross-functionally with Product, Data Science, Infrastructure, and Engineering teams to translate complex problems into ML solutions</li>\n<li>Improve system performance across latency, throughput, and model quality metrics</li>\n<li>Research and apply state-of-the-art machine learning and AI techniques, including deep learning, graph &amp; transformers based, and LLM evaluation/alignment</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>3-5+ years of experience building, deploying, and operating machine learning systems in production</li>\n<li>Strong programming skills in Python, Java, Go, or similar languages, with solid software engineering fundamentals</li>\n<li>ML Fundamentals: a strong grasp of algorithms, from classic statistical learning (XGBoost, Random Forests, regressions) to DL architectures (Transformers, CNNs, GNNs)</li>\n<li>Hands-on experience with modern ML frameworks (e.g., PyTorch, TensorFlow)</li>\n<li>Experience designing scalable ML pipelines, data processing systems, and model serving infrastructure</li>\n<li>Ability to work cross-functionally and translate ambiguous product or business problems into technical solutions</li>\n<li>Experience improving measurable metrics through applied machine learning</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with recommender systems, search/ranking systems, advertising/auction systems, large-scale representation learning, or multimodal embedding systems</li>\n<li>Familiarity with distributed systems and large-scale data processing (Spark, Kafka, Ray, Airflow, BigQuery, Redis, etc.)</li>\n<li>Experience working with real-time systems and low-latency production environments</li>\n<li>Background in feature engineering, model optimization, and production monitoring</li>\n<li>Experience with LLM/Gen AI techniques, including but not limited to LLM evaluation, alignment, fine-tuning, knowledge distillation, RAG/agentic systems and productionizing LLM-powered products at scale</li>\n<li>Advanced degree in Computer Science, Machine Learning, or related quantitative field</li>\n</ul>\n<p>Potential Teams:</p>\n<ul>\n<li>Ads Measurement Modeling</li>\n<li>Ads Targeting and Retrieval</li>\n<li>Advertiser Optimization</li>\n<li>Ads Marketplace Quality</li>\n<li>Ads Creative Effectiveness</li>\n<li>Ads Foundational Representations</li>\n<li>Ads Content Understanding</li>\n<li>Ads Ranking</li>\n<li>Feed Relevance</li>\n<li>Search and Answers Relevance</li>\n<li>ML Understanding</li>\n<li>Notifications Relevance</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Comprehensive Healthcare Benefits and Income Replacement Programs</li>\n<li>401k with Employer Match</li>\n<li>Global Benefit programs that fit your lifestyle, from workspace to professional development to caregiving support</li>\n<li>Family Planning Support</li>\n<li>Gender-Affirming Care</li>\n<li>Mental Health &amp; Coaching Benefits</li>\n<li>Flexible Vacation &amp; Paid Volunteer Time Off</li>\n<li>Generous Paid Parental Leave</li>\n</ul>\n<p>Pay Transparency:</p>\n<p>This job posting may span more than one career level. In addition to base salary, this job is eligible to receive equity in the form of restricted stock units, and depending on the position offered, it may also be eligible to receive a commission. Additionally, Reddit offers a wide range of benefits to U.S.-based employees, including medical, dental, and vision insurance, 401(k) program with employer match, generous time off for vacation, and parental leave.</p>\n<p>To provide greater transparency to candidates, we share base salary ranges for all US-based job postings regardless of state. We set standard base pay ranges for all roles based on function, level, and country location, benchmarked against similar stage growth companies. Final offer amounts are determined by multiple factors including, skills, depth of work experience and relevant licenses/credentials, and may vary from the amounts listed below.</p>\n<p>The base salary range for this position is: $185,800-$260,100 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_467be5c4-940","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7131932","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$185,800-$260,100 USD","x-skills-required":["Python","Java","Go","PyTorch","TensorFlow","XGBoost","Random Forests","Regressions","Transformers","CNNs","GNNs","Spark","Kafka","Ray","Airflow","BigQuery","Redis"],"x-skills-preferred":["Recommender systems","Search/ranking systems","Advertising/auction systems","Large-scale representation learning","Multimodal embedding systems","Distributed systems","Large-scale data processing","Real-time systems","Low-latency production environments","Feature engineering","Model optimization","Production monitoring","LLM/Gen AI techniques"],"datePosted":"2026-04-18T15:57:49.850Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Go, PyTorch, TensorFlow, XGBoost, Random Forests, Regressions, Transformers, CNNs, GNNs, Spark, Kafka, Ray, Airflow, BigQuery, Redis, Recommender systems, Search/ranking systems, Advertising/auction systems, Large-scale representation learning, Multimodal embedding systems, Distributed systems, Large-scale data processing, Real-time systems, Low-latency production environments, Feature engineering, Model optimization, Production monitoring, LLM/Gen AI techniques","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":185800,"maxValue":260100,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_03224784-9c2"},"title":"Senior Data Engineering Manager","description":"<p>Job Title: Senior Data Engineering Manager</p>\n<p>Location: Dublin, Ireland</p>\n<p>Department: R&amp;D</p>\n<p>Job Description:</p>\n<p>Intercom is seeking a Senior Data Engineering Manager to lead the design and evolution of the core infrastructure that powers our entire data ecosystem. As a leader, you will partner with product and business teams to drive key data initiatives and ensure the success of our data engineering team.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Next-Gen Platform Evolution: Partner with product and business teams to design and implement the next generation of our data stack, ensuring it can meet the demands of advanced analytics and AI applications.</li>\n</ul>\n<ul>\n<li>Enablement Through Tooling: Partner closely with Analytics Engineers, Analysts, and Data Scientists to build self-service tooling and infrastructure that enables them to move fast and deploy safely.</li>\n</ul>\n<ul>\n<li>Data Quality Guardianship: Implement advanced monitoring systems to proactively detect, surface, and resolve data quality issues across our high-throughput environment.</li>\n</ul>\n<ul>\n<li>Driving Automation: Develop automation and tooling that streamlines the creation and discovery of high-quality analytics data, making the entire data lifecycle more efficient.</li>\n</ul>\n<p>Strategic Impact You&#39;ll Drive:</p>\n<ul>\n<li>GTM Data Platform Strategy: Build the data acquisition strategy that will enable us to build the next generation of business-focused internal software.</li>\n</ul>\n<ul>\n<li>Conversational BI Strategy: Lead the charge to shift away from complex, technical reporting toward natural language interaction to make data truly democratized and accessible.</li>\n</ul>\n<ul>\n<li>Platform &amp; Warehousing Strategy: Lead the architectural- and cost review and revamp of our core data infrastructure to ensure it can scale exponentially for future growth and advanced use cases.</li>\n</ul>\n<p>Recent Wins You&#39;ll Build Upon:</p>\n<ul>\n<li>AI-assisted Local Analytics Development Environment for Airflow and DBT.</li>\n</ul>\n<ul>\n<li>Data-rich AI apps containerized on Snowflake SPCS.</li>\n</ul>\n<ul>\n<li>A new, modern data catalog solution.</li>\n</ul>\n<ul>\n<li>Migrating critical MySQL ingestion pipelines from Aurora to PlanetScale.</li>\n</ul>\n<p>Who You Are:</p>\n<ul>\n<li>A leader, a builder, and a problem-solver who thrives on solving real-world business problems.</li>\n</ul>\n<ul>\n<li>7+ years of experience in the data space, leading teams of 6+ engineers.</li>\n</ul>\n<ul>\n<li>Stakeholder focus: ability to communicate complex technical solutions to a business-focused audience and vice versa.</li>\n</ul>\n<ul>\n<li>Technical depth: not afraid to get hands dirty and write code when needed.</li>\n</ul>\n<ul>\n<li>A leader and mentor: naturally recognizes opportunities to step back and mentor others.</li>\n</ul>\n<p>Bonus Points (Our Modern Stack Knowledge):</p>\n<ul>\n<li>Airflow at scale: extensive experience working with Apache Airflow, especially the nuances of operating it reliably in a high-volume environment.</li>\n</ul>\n<ul>\n<li>Modern data stack fluency: familiarity with tools like Snowflake and DBT.</li>\n</ul>\n<ul>\n<li>Future-focused: keeps a keen eye on industry trends and emerging technologies.</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Competitive salary and equity in a fast-growing start-up.</li>\n</ul>\n<ul>\n<li>We serve lunch every weekday, plus a variety of snack foods and a fully stocked kitchen.</li>\n</ul>\n<ul>\n<li>Regular compensation reviews - we reward great work!</li>\n</ul>\n<ul>\n<li>Pension scheme &amp; match up to 4%.</li>\n</ul>\n<ul>\n<li>Peace of mind with life assurance, as well as comprehensive health and dental insurance for you and your dependents.</li>\n</ul>\n<ul>\n<li>Open vacation policy and flexible holidays so you can take time off when you need it.</li>\n</ul>\n<ul>\n<li>Paid maternity leave, as well as 6 weeks paternity leave for fathers, to let you spend valuable time with your loved ones.</li>\n</ul>\n<ul>\n<li>If you’re cycling, we’ve got you covered on the Cycle-to-Work Scheme. With secure bike storage too.</li>\n</ul>\n<ul>\n<li>MacBooks are our standard, but we also offer Windows for certain roles when needed.</li>\n</ul>\n<p>Policies:</p>\n<ul>\n<li>Intercom has a hybrid working policy. We believe that working in person helps us stay connected, collaborate easier and create a great culture while still providing flexibility to work from home.</li>\n</ul>\n<ul>\n<li>We have a radically open and accepting culture at Intercom. We avoid spending time on divisive subjects to foster a safe and cohesive work environment for everyone.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_03224784-9c2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Intercom","sameAs":"https://www.intercom.com/","logo":"https://logos.yubhub.co/intercom.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/intercom/jobs/7574762","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Apache Airflow","DBT","Snowflake","Data Engineering","Data Science","Analytics","Data Management","Data Quality","Automation","Cloud Computing","Data Warehousing","Big Data","Machine Learning","Artificial Intelligence"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:06.635Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, Ireland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Airflow, Apache Airflow, DBT, Snowflake, Data Engineering, Data Science, Analytics, Data Management, Data Quality, Automation, Cloud Computing, Data Warehousing, Big Data, Machine Learning, Artificial Intelligence"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0a154c39-08a"},"title":"Senior Machine Learning Platform Engineer (Platform)","description":"<p>Ready to be pushed beyond what you think you’re capable of?</p>\n<p>At Coinbase, our mission is to increase economic freedom in the world.</p>\n<p>We&#39;re seeking a Senior Machine Learning Platform Engineer to join our Machine Learning Platform team. The team builds the foundational components for feature engineering and training/serving ML models at Coinbase. Our platform is used to combat fraud, personalize user experiences, and to analyze blockchains.</p>\n<p>As a Senior Machine Learning Platform Engineer, you will:</p>\n<p>Form a deep understanding of our Machine Learning Engineers’ needs and our current capabilities and gaps. Mentor our talented junior engineers on how to build high quality software, and take their skills to the next level. Continually raise our engineering standards to maintain high-availability and low-latency for our ML inference infrastructure that runs both predictive ML models and LLMs. Optimize low latency streaming pipelines to give our ML models the freshest and highest quality data. Evangelize state-of-the-art practices on building high-performance distributed training jobs that process large volumes of data. Build tooling to observe the quality of data going into our models and to detect degradations impacting model performance.</p>\n<p>What we look for in you:</p>\n<p>5+ yrs of industry experience as a Software Engineer. Strong understanding of distributed systems. Lead by example through high quality code and excellent communication skills. Great sense of design, and can bring clarity to complex technical requirements. Treat other engineers as a customer, and have an obsessive focus on delivering them a seamless experience. Mastery of the fundamentals, such that you can quickly jump between many varied technologies and still operate at a high level. Demonstrates the ability to responsibly use generative AI tools and copilots (e.g., LibreChat, Gemini, Glean) in daily workflows, continuously learn as tools evolve, and apply human-in-the-loop practices to deliver business-ready outputs and drive measurable improvements in efficiency, cost, and quality.</p>\n<p>Nice to haves:</p>\n<p>Experience building ML models and working with ML systems. Experience working on a platform team, and building developer tooling. Experience with the technologies we use (Python, Golang, Ray, Tecton, Spark, Airflow, Databricks, Snowflake, and DynamoDB).</p>\n<p>Job ID: P75535</p>\n<p>Pay Transparency Notice: Depending on your work location, the target annual base salary for this position can range as detailed below. Total compensation may also include equity and bonus eligibility and benefits (including medical, dental, vision and 401(k)). Annual base salary range (excluding equity and bonus): $186,065-$225,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0a154c39-08a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Coinbase","sameAs":"https://www.coinbase.com/","logo":"https://logos.yubhub.co/coinbase.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coinbase/jobs/7604203","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$186,065-$225,000 USD","x-skills-required":["distributed systems","high-quality code","excellent communication skills","design","fundamentals","generative AI tools","copilots"],"x-skills-preferred":["ML models","ML systems","platform team","developer tooling","Python","Golang","Ray","Tecton","Spark","Airflow","Databricks","Snowflake","DynamoDB"],"datePosted":"2026-04-18T15:56:24.447Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"distributed systems, high-quality code, excellent communication skills, design, fundamentals, generative AI tools, copilots, ML models, ML systems, platform team, developer tooling, Python, Golang, Ray, Tecton, Spark, Airflow, Databricks, Snowflake, DynamoDB","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":186065,"maxValue":225000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1f2f48ad-46d"},"title":"Senior Analytics Engineer","description":"<p>We&#39;re looking for a dedicated Analytics Engineer to join the AI Group to help us with data platform development, cross-functional collaboration, data strategy &amp; governance, advanced analytics &amp; insights, automation &amp; optimization, innovation in data infrastructure, and strategic influence.</p>\n<p>As an Analytics Engineer, you will design, build, and manage scalable data pipelines and ETL processes to support a robust, analytics-ready data platform. You will partner with AI analysts, ML scientists, engineers, and business teams to understand data needs and ensure accurate, reliable, and ergonomic data solutions. You will lead initiatives in data model development, data quality ownership, warehouse management, and production support for critical workflows. You will conduct data analysis and build custom models to support strategic business decisions and performance measurement. You will streamline data collection and reporting processes to reduce manual effort and improve efficiency. You will create scalable solutions like unified data pipelines and access control systems to meet evolving organisational needs. You will work with partner teams to align data collection with long-term analytics and feature development goals.</p>\n<p>We&#39;re looking for someone who writes advanced SQL with a preference for well-architected data models, optimized query performance, and clearly documented code. You should be familiar with the modern data stack, including dbt and Snowflake. You should have a growth mindset and eagerness to learn. You should exhibit great judgment and sharp business and product instincts that allow you to differentiate essential versus nice-to-have and to make good choices about trade-offs. You should practice excellent communication skills, and you should tailor explanations of technical concepts to a variety of audiences.</p>\n<p>Nice to have: exposure to Apache Airflow or other DAG frameworks, worked in Tableau, Looker, or similar visualization/business intelligence platform, experience with operational tools and business systems like Google Analytics, Marketo, Salesforce, Segment, or Stripe, familiarity with Python.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1f2f48ad-46d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Intercom","sameAs":"https://www.intercom.com/","logo":"https://logos.yubhub.co/intercom.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/intercom/jobs/7807847","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["advanced SQL","dbt","Snowflake","data pipeline development","ETL process management","data strategy & governance","advanced analytics & insights","automation & optimization","innovation in data infrastructure","strategic influence"],"x-skills-preferred":["Apache Airflow","Tableau","Looker","Google Analytics","Marketo","Salesforce","Segment","Stripe","Python"],"datePosted":"2026-04-18T15:55:10.503Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Dublin, Ireland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"advanced SQL, dbt, Snowflake, data pipeline development, ETL process management, data strategy & governance, advanced analytics & insights, automation & optimization, innovation in data infrastructure, strategic influence, Apache Airflow, Tableau, Looker, Google Analytics, Marketo, Salesforce, Segment, Stripe, Python"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9be280f4-cbc"},"title":"Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for an engineer to join our small, high-impact team responsible for architecting and scaling the core infrastructure behind distributed training pipelines, multimodal data catalogs, and intelligent processing systems that operate over petabytes of data.</p>\n<p>As a software engineer on our data infrastructure team, you&#39;ll design, build, and operate scalable, fault-tolerant infrastructure for LLM Research: distributed compute, data orchestration, and storage across modalities. You&#39;ll develop high-throughput systems for data ingestion, processing, and transformation , including training data catalogs, deduplication, quality checks, and search. You&#39;ll also build systems for traceability, reproducibility, and robust quality control at every stage of the data lifecycle.</p>\n<p>You&#39;ll collaborate with research teams to unlock new features, improve data quality, and accelerate training cycles. You&#39;ll implement and maintain monitoring and alerting to support platform reliability and performance.</p>\n<p>If you&#39;re excited by distributed systems, large-scale data mining, open-source tools like Spark, Kafka, Beam, Ray, and Delta Lake, and enjoy building from the ground up, we&#39;d love to hear from you.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9be280f4-cbc","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Thinking Machines Lab","sameAs":"https://thinkingmachines.ai/","logo":"https://logos.yubhub.co/thinkingmachines.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/thinkingmachines/jobs/5013919008","x-work-arrangement":"onsite","x-experience-level":"entry|mid|senior","x-job-type":"full-time","x-salary-range":"$350,000 - $475,000 USD","x-skills-required":["backend language (Python or Rust)","distributed compute frameworks (Apache Spark or Ray)","cloud infrastructure","data lake architectures","batch and streaming pipelines"],"x-skills-preferred":["Kafka","dbt","Terraform","Airflow","web crawler","deduplication","data mining","search","file formats and storage systems"],"datePosted":"2026-04-18T15:54:00.309Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend language (Python or Rust), distributed compute frameworks (Apache Spark or Ray), cloud infrastructure, data lake architectures, batch and streaming pipelines, Kafka, dbt, Terraform, Airflow, web crawler, deduplication, data mining, search, file formats and storage systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":350000,"maxValue":475000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1aad838f-387"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability.</p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li>Data Governance &amp; Access Control: Design and implement robust access control systems ensuring only authorized users can access sensitive data.</li>\n<li>Financial Data Infrastructure: Build and maintain data pipelines and warehouses powering business-critical reporting.</li>\n<li>Cloud Storage &amp; Reliability: Architect disaster recovery, backup, and replication systems for petabyte-scale data.</li>\n<li>Data Platform &amp; Tooling: Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark.</li>\n</ul>\n<p>You&#39;ll work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p>To be successful in this role, you&#39;ll need:</p>\n<ul>\n<li>10+ years of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems.</li>\n<li>3+ years of experience leading large scale, complex projects or teams as an engineer or tech lead.</li>\n<li>Deep experience with at least one of:</li>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar.</li>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS).</li>\n<li>Can navigate complex technical tradeoffs between performance, cost, security, and maintainability.</li>\n<li>Have excellent collaboration skills - you work well with both technical and non-technical stakeholders.</li>\n</ul>\n<p>Strong candidates may also have:</p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure.</li>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures.</li>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale.</li>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks.</li>\n<li>Experience working in fintech, financial services, or highly regulated environments.</li>\n<li>Security engineering background with focus on data protection and access controls.</li>\n</ul>\n<p>Technologies We Use:</p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran.</li>\n<li>Storage: GCS, S3.</li>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS.</li>\n<li>Languages: Python, Go, SQL.</li>\n</ul>\n<p>The annual compensation range for this role is $405,000-$485,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1aad838f-387","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000-$485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures"],"x-skills-preferred":["data warehousing","ETL/ELT pipelines","analytics infrastructure","data reliability","availability","cost efficiency","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"datePosted":"2026-04-18T15:52:47.297Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, data reliability, availability, cost efficiency, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ba30b234-c68"},"title":"Senior Data Engineer, Payments","description":"<p>We&#39;re looking for a Senior Data Engineer to join our Payments team. As a critical part of our operations, you&#39;ll handle data related to compliance with Tax, Payments, and Legal regulations. You&#39;ll design, build, and maintain robust and efficient data pipelines that collect, process, and store data from various sources, including user interactions, listing details, and external data feeds.</p>\n<p>Your work will involve developing data models that enable the efficient analysis and manipulation of data for merchandising optimization, ensuring data quality, consistency, and accuracy. You&#39;ll also develop high-quality data assets for product use-cases by partnering with Product, AI/ML, and Data Science teams.</p>\n<p>As a Senior Data Engineer, you&#39;ll contribute to creating standards and best practices for Airbnb&#39;s Data Engineering and shape the tools, processes, and standards used by the broader data community. You&#39;ll collaborate with cross-functional teams to define data requirements and deliver data solutions that drive merchandising and sales improvements.</p>\n<p>To succeed in this role, you&#39;ll need 6+ years of relevant industry experience, a BE/B.Tech in Computer Science or a relevant technical degree, and hands-on experience in DSA coding, data structure, and algorithm. You&#39;ll also need extensive experience designing, building, and operating robust distributed data platforms and handling data at the petabyte scale.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ba30b234-c68","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7256787","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Scala","Python","data processing technologies","query authoring (SQL)","ETL schedulers (Apache Airflow, Luigi, Oozie, AWS Glue)","data warehousing concepts","relational databases (PostgreSQL, MySQL)","columnar databases (Redshift, BigQuery, HBase, ClickHouse)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:13.348Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Scala, Python, data processing technologies, query authoring (SQL), ETL schedulers (Apache Airflow, Luigi, Oozie, AWS Glue), data warehousing concepts, relational databases (PostgreSQL, MySQL), columnar databases (Redshift, BigQuery, HBase, ClickHouse)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1e09f714-7db"},"title":"Analytics Engineer, FinTech","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world&#39;s largest networks that powers millions of websites and other internet properties, from individual bloggers to Fortune 500 companies, protecting and accelerating them without adding hardware, installing software, or changing a line of code.</p>\n<p>Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. Cloudflare was named to Entrepreneur Magazine&#39;s Top Company Cultures list and ranked among the World&#39;s Most Innovative Companies by Fast Company.</p>\n<p>The FinTech Data Science team is central to Cloudflare&#39;s innovation and harnesses the massive amount of data generated by our network. We cover a broad scope, from optimizing Billing and Revenue operations to detecting Fraud, and possess a unique opportunity to use these insights to discover new products or transform existing ones.</p>\n<p>About the Role</p>\n<p>We are looking for an Analytics Engineer to join our FinTech Data Science team who cares deeply about data quality and usability. Sitting at the intersection of data engineering and analysis, you will be the architect of our data layer. While our Data Scientists focus on automating decisions, you will focus on the &#39;truth&#39; of the data , ensuring that the tables and dashboards powering our decisions are accurate, accessible, documented, and reliable.</p>\n<p>You will transform raw tables into canonical data models and own the presentation layer that leadership uses to monitor the health of our business. If you are excited to build the foundational data infrastructure that powers a multi-billion dollar fintech operation, we would love to hear from you!</p>\n<p>Day-to-day responsibilities include:</p>\n<ul>\n<li>Build out the canonical data schema for FinTech and related organizations by designing and maintaining well-structured, modular, and user-friendly data tables.</li>\n</ul>\n<ul>\n<li>Design, develop, deploy, and operate high-quality production ELT pipelines and data architectures, integrating data from various sources and formats.</li>\n</ul>\n<ul>\n<li>Architect and maintain the presentation layer in BI tools (e.g., Looker/Superset) to ensure dashboards are performant and provide a seamless self-serve experience.</li>\n</ul>\n<ul>\n<li>Act as a strategic partner to stakeholders by translating vague business questions into concrete technical solutions that drive business value.</li>\n</ul>\n<ul>\n<li>Ensure data is accurate, complete, and timely by implementing robust testing, monitoring, and validation protocols for your code and data.</li>\n</ul>\n<ul>\n<li>Establish and share best practices in performance, code quality, data governance, and discoverability while participating in mentoring initiatives.</li>\n</ul>\n<p>Required skills, knowledge, and experience:</p>\n<ul>\n<li>5+ years of experience in Analytics Engineering, Data Engineering, or related roles working with big data at scale.</li>\n</ul>\n<ul>\n<li>Expert-level SQL and proficiency in a high-level scripting language (e.g., Python, R, or Scala) for data automation and manipulation.</li>\n</ul>\n<ul>\n<li>Experience with workflow management tools (e.g., Airflow) to schedule and monitor complex data pipelines.</li>\n</ul>\n<ul>\n<li>Strong experience with dbt or similar frameworks for transforming data in the warehouse.</li>\n</ul>\n<ul>\n<li>Deep experience with BI tools (e.g., Looker, Superset, or Grafana) and a strong understanding of how to structure data for downstream consumption.</li>\n</ul>\n<ul>\n<li>Solid foundation in software best practices, including version control (Git), CI/CD, and data testing/quality frameworks.</li>\n</ul>\n<ul>\n<li>Ability to operate comfortably in a fast-paced environment and take ownership of projects with minimal oversight.</li>\n</ul>\n<ul>\n<li>Excellent communication skills with the ability to bridge the gap between technical engineering terms and business requirements.</li>\n</ul>\n<ul>\n<li>A learning mindset and exceptional curiosity,eagerly diving into new domains and bringing informed ideas to the table.</li>\n</ul>\n<p>Bonus Points</p>\n<p>Experience in FinTech</p>\n<p>What Makes Cloudflare Special?</p>\n<p>We&#39;re not just a highly ambitious, large-scale technology company. We&#39;re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare&#39;s enterprise customers,at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use,it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal,we don&#39;t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1e09f714-7db","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7649684","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Python","R","Scala","Airflow","dbt","Looker","Superset","Grafana","Git","CI/CD","data testing/quality frameworks"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:02.907Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, R, Scala, Airflow, dbt, Looker, Superset, Grafana, Git, CI/CD, data testing/quality frameworks"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_fa9a54d7-549"},"title":"Senior Site Reliability Engineer, Data Infrastructure","description":"<p>As a Senior Site Reliability Engineer, you will own the reliability and performance of our Kubernetes-based data platform. You will design and operate highly available, multi-region systems, ensuring our services meet strict uptime and latency targets.</p>\n<p>Day-to-day, you’ll work on scaling infrastructure, improving deployment pipelines, and hardening our security posture. You’ll play a key role in evolving our DevSecOps practices while partnering closely with engineering teams to ensure services are built for reliability from day one.</p>\n<p>We operate with production-grade discipline, supporting mission-critical services with stringent uptime requirements and a focus on automation, observability, and resilience.</p>\n<p>The Platform &amp; Infrastructure Engineering team in the Data Infrastructure organization is responsible for the reliability, scalability, and security of the company’s data platform. The team builds and operates the foundational systems that power data ingestion, transformation, analytics, and internal AI workloads at scale.</p>\n<p>About the role:</p>\n<ul>\n<li>5+ years of experience in Site Reliability Engineering, Platform Engineering, or Infrastructure Engineering roles</li>\n<li>Deep expertise in Kubernetes and containerized software services, including cluster design, operations, and troubleshooting in production environments</li>\n<li>Strong experience building and operating CI/CD systems, including tools such as Argo CD and GitHub Actions</li>\n<li>Proven experience owning production systems with high availability requirements (≥99.99% uptime), including incident response, SLI/SLO/SLA definition, error budgets, and postmortems</li>\n<li>Hands-on experience designing and operating geo-replicated, multi-region, active-active systems, including traffic routing, failover strategies, and data consistency tradeoffs</li>\n<li>Strong experience building and owning observability components, including metrics, logging, and tracing (e.g., Prometheus, Grafana, OpenTelemetry).</li>\n<li>Experience with infrastructure as code (e.g., Helm, Terraform, Pulumi) and automated environment provisioning</li>\n<li>Strong understanding of system performance tuning, capacity planning, and resource optimization in distributed systems</li>\n<li>Experience implementing and operating security best practices in cloud-native environments (e.g., secrets management, network policies, vulnerability scanning)</li>\n</ul>\n<p>Preferred:</p>\n<ul>\n<li>Experience operating data platforms or data-intensive workloads (e.g., Spark, Airflow, Kafka, Flink)</li>\n<li>Familiarity with service mesh technologies (e.g., Istio, Linkerd)</li>\n<li>Experience working in regulated environments with compliance frameworks such as GDPR, SOC 2, HIPAA, or SOX</li>\n<li>Background in building internal developer platforms or self-service infrastructure</li>\n</ul>\n<p>Wondering if you’re a good fit?</p>\n<p>We believe in investing in our people, and value candidates who can bring their own diversified experiences to our teams – even if you aren’t a 100% skill or experience match.</p>\n<p>Here are a few qualities we’ve found compatible with our team. If some of this describes you, we’d love to talk.</p>\n<ul>\n<li>You love building highly reliable systems that operate at scale</li>\n<li>You’re curious about how to continuously improve system resilience, security, and operations</li>\n<li>You’re an expert in diagnosing and solving complex distributed systems problems</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We’re in an exciting stage of hyper-growth that you will not want to miss out on. We’re not afraid of a little chaos, and we’re constantly learning.</p>\n<p>Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems.</p>\n<p>As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too.</p>\n<p>Come join us!</p>\n<p>The base salary range for this role is $165,000 to $242,000. The starting salary will be determined based on job-related knowledge, skills, experience, and market location. We strive for both market alignment and internal equity when determining compensation.</p>\n<p>In addition to base salary, our total rewards package includes a discretionary bonus, equity awards, and a comprehensive benefits program (all based on eligibility).</p>\n<p>What We Offer</p>\n<p>The range we’ve posted represents the typical compensation range for this role. To determine actual compensation, we review the market rate for each candidate which can include a variety of factors. These include qualifications, experience, interview performance, and location.</p>\n<p>In addition to a competitive salary, we offer a variety of benefits to support your needs, including:</p>\n<ul>\n<li>Medical, dental, and vision insurance</li>\n<li>100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Our Workplace</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets.</p>\n<p>New hires will be invited to attend onboarding at one of our hubs within their first month.</p>\n<p>Teams also gather quarterly to support collaboration.</p>\n<p>California Consumer Privacy Act - California applicants only</p>\n<p>CoreWeave is an equal opportunity employer, committed to fostering an inclusive and supportive workplace.</p>\n<p>All qualified applicants and candidates will receive consideration for employment without regard to race, color, religion, sex, disability, age, sexual orientation, gender identity, national origin, veteran status, or genetic information.</p>\n<p>As part of this commitment and consistent with the Americans with Disabilities Act (ADA), CoreWeave will ensure that qualified applicants and candidates with disabilities are provided reasonable accommodations for the hiring process, unless such accommodation would cause an undue hardship.</p>\n<p>If reasonable accommodation is needed, please contact: careers@coreweave.com.</p>\n<p>Export Control Compliance</p>\n<p>This position requires access to export controlled information.</p>\n<p>To conform to U.S. Government export regulations applicable to that information, applicant must either be (A) a U.S. person, defined as a (i) U.S. citizen or national, (ii) U.S. lawful permanent resident (green card holder), (iii) refugee under 8 U.S.C. § 1157, or (iv) asylee under 8 U.S.C. § 1158, (B) eligible to access the export controlled information without restrictions, or (C) otherwise exempt from the export regulations.</p>\n<p>If you are not a U.S. person, you will be required to provide documentation of your eligibility to access the export controlled information before being considered for this position.</p>\n<p>Please note that CoreWeave is subject to the requirements of the U.S. Department of Commerce&#39;s Export Administration Regulations (EAR) and the U.S. Department of State&#39;s International Traffic in Arms Regulations (ITAR).</p>\n<p>By applying for this position, you acknowledge that you have read and understood the export control requirements and that you will comply with them.</p>\n<p>If you have any questions or concerns regarding the export control requirements, please contact: careers@coreweave.com.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_fa9a54d7-549","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4671535006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$165,000 to $242,000","x-skills-required":["Kubernetes","containerized software services","cluster design","operations","troubleshooting","CI/CD systems","Argo CD","GitHub Actions","production systems","high availability","incident response","SLI/SLO/SLA definition","error budgets","postmortems","geo-replicated","multi-region","active-active systems","traffic routing","failover strategies","data consistency tradeoffs","observability components","metrics","logging","tracing","Prometheus","Grafana","OpenTelemetry","infrastructure as code","Helm","Terraform","Pulumi","automated environment provisioning","system performance tuning","capacity planning","resource optimization","distributed systems","security best practices","cloud-native environments","secrets management","network policies","vulnerability scanning"],"x-skills-preferred":["Spark","Airflow","Kafka","Flink","service mesh technologies","Istio","Linkerd","regulated environments","compliance frameworks","GDPR","SOC 2","HIPAA","SOX","internal developer platforms","self-service infrastructure"],"datePosted":"2026-04-18T15:51:59.035Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, NY / Bellevue, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, containerized software services, cluster design, operations, troubleshooting, CI/CD systems, Argo CD, GitHub Actions, production systems, high availability, incident response, SLI/SLO/SLA definition, error budgets, postmortems, geo-replicated, multi-region, active-active systems, traffic routing, failover strategies, data consistency tradeoffs, observability components, metrics, logging, tracing, Prometheus, Grafana, OpenTelemetry, infrastructure as code, Helm, Terraform, Pulumi, automated environment provisioning, system performance tuning, capacity planning, resource optimization, distributed systems, security best practices, cloud-native environments, secrets management, network policies, vulnerability scanning, Spark, Airflow, Kafka, Flink, service mesh technologies, Istio, Linkerd, regulated environments, compliance frameworks, GDPR, SOC 2, HIPAA, SOX, internal developer platforms, self-service infrastructure","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":165000,"maxValue":242000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1be89b3c-bc1"},"title":"Staff Analytics Engineer","description":"<p>We are currently hiring for multiple teams:</p>\n<p>Foundational Data team: Our mission in the Foundational Data team is to build and maintain high-quality datasets frequently used across all of Airbnb. We set company-wide standards that decide how locations are grouped into regions, visitors are measured based upon site traffic, bot traffic is separated from organic traffic, and cloud costs are attributed to Airbnb services. This data is used to build public financial reports, drive strategic marketing decisions, and manage operational costs.</p>\n<p>AirCover Data Foundation: The AirCover Data Foundation team is responsible for providing trustworthy, consistent data and metrics to facilitate business insights, informed decision-making, and seamless operations across Airbnb&#39;s AirCover programs, such as Guest Travel Insurance, AirCover for Hosts, and AirCover for Guests.</p>\n<p>As a Staff Analytics Engineer, you will bring a unique lens to our data strategy and provide in-depth technical mentorship and leadership to the team. We are looking for someone with expertise in data modeling, metric development, and large-scale distributed data processing frameworks like Presto or Spark.</p>\n<p>Leveraging our internal, top-tier data tooling alongside other resources, you will empower both technical and non-technical teams across Airbnb to utilize our data for making decisions grounded in evidence. Staff-level engineers are expected to do this with a minimal amount of supervision. We value innovative thinkers who consistently seek smarter and more efficient solutions while managing daily operations, deadlines, and collaborating with team members.</p>\n<p>A Typical Day:</p>\n<ul>\n<li>Develop high-quality data assets to satisfy a wide range of use-cases</li>\n<li>Develop frameworks and tools to scale insight generation to meet critical business and infrastructure requirements</li>\n<li>Collaborate and build strong partnerships with other data practitioners throughout Airbnb</li>\n<li>Influence the trajectory of data in decision making</li>\n<li>Improve trust in our data by championing for data quality across the stack</li>\n</ul>\n<p>Your Expertise:</p>\n<ul>\n<li>9+ years of experience with a BS/Masters or 6+ years with a PhD</li>\n<li>Fluent in SQL and proficient in at least one data engineering language, such as Python or Scala</li>\n<li>Expertise using business intelligence and reporting tools like Superset and Tableau</li>\n<li>Expertise in large-scale distributed data processing frameworks like Presto or Spark</li>\n<li>Expertise in data modeling for data warehouses and/or metrics repositories</li>\n<li>Experience with an ETL framework like Airflow</li>\n<li>Clear and mature communication skills: ability to distill complex ideas for technical and non-technical stakeholders</li>\n<li>Ability to provide technical leadership and mentorship, guiding teams on best practices and contributing to the development of analytic engineering strategies</li>\n<li>Experience exploring and leveraging LLM AI’s in everyday tasks (coding, documentation, etc…)</li>\n<li>Strong capability to forge trusted partnerships across working teams</li>\n</ul>\n<p>Nice to have:</p>\n<ul>\n<li>Scaling data tasks via automation</li>\n<li>Previous experience in large-scale cloud-based software engineering or system architecture</li>\n<li>Experience with AB experimentation</li>\n<li>Familiarity with AI/ML algorithms, including their dependencies on data, as well as their respective strengths and limitations</li>\n<li>Designing and/or leveraging high-quality data visualization tools</li>\n</ul>\n<p>Your Location: This position is US - Remote Eligible. The role may include occasional work at an Airbnb office or attendance at offsites, as agreed to with your manager. While the position is Remote Eligible, you must live in a state where Airbnb, Inc. has a registered entity. Click here for the up-to-date list of excluded states.</p>\n<p>Our Commitment To Inclusion &amp; Belonging: Airbnb is committed to working with the broadest talent pool possible. We believe diverse ideas foster innovation and engagement, and allow us to attract creatively-led people, and to develop the best products, services and solutions. All qualified individuals are encouraged to apply. We strive to also provide a disability inclusive application and interview process. If you are a candidate with a disability and require reasonable accommodation in order to submit an application, please contact us at: reasonableaccommodations@airbnb.com.</p>\n<p>How We&#39;ll Take Care of You: Our job titles may span more than one career level. The actual base pay is dependent upon many factors, such as: training, transferable skills, work experience, business needs and market demands. The base pay range is subject to change and may be modified in the future. This role may also be eligible for bonus, equity, benefits, and Employee Travel Credits. Pay Range $194,000-$240,000 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1be89b3c-bc1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7733495","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$194,000-$240,000 USD","x-skills-required":["SQL","Python","Scala","Presto","Spark","Superset","Tableau","ETL","Airflow","Data Modeling","Data Warehousing","Metrics Repositories"],"x-skills-preferred":["LLM AI","AI/ML Algorithms","Data Visualization","Cloud-Based Software Engineering","System Architecture","AB Experimentation"],"datePosted":"2026-04-18T15:50:40.547Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Scala, Presto, Spark, Superset, Tableau, ETL, Airflow, Data Modeling, Data Warehousing, Metrics Repositories, LLM AI, AI/ML Algorithms, Data Visualization, Cloud-Based Software Engineering, System Architecture, AB Experimentation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":194000,"maxValue":240000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_059293a1-afa"},"title":"Systems Engineer, Data","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>We were named to Entrepreneur Magazine’s Top Company Cultures list and ranked among the World’s Most Innovative Companies by Fast Company.</p>\n<p>About the Team</p>\n<p>The Core Data team’s mission is building a centralized data platform for Cloudflare that provides secure, democratized access to data for internal customers throughout the company. We operate infrastructure and craft tools to empower both technical and non-technical users to answer their most important questions. We facilitate access to data from federated sources across the company for dashboarding, ad-hoc querying and in-product use cases. We power data pipelines and data products, secure and monitor data, and drive data governance at Cloudflare.</p>\n<p>Our work enables every individual at the company to act with greater information and make more informed decisions.</p>\n<p>About the Role</p>\n<p>We are looking for a systems engineer with a strong background in data to help us expand and maintain our data infrastructure. You’ll contribute to the technical implementation of our scaling data platform, manage access while accounting for privacy and security, build data pipelines, and develop tools to automate accessibility and usefulness of data. You’ll collaborate with teams including Product Growth, Marketing, and Billing to help them make informed decisions and power usage-based invoicing platforms, as well as work with product teams to bring new data-driven solutions to Cloudflare customers.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Contribute to the design and execution of technical architecture for highly visible data infrastructure at the company.</li>\n<li>Design and develop tools and infrastructure to improve and scale our data systems at Cloudflare.</li>\n<li>Build and maintain data pipelines and data products to serve customers throughout the company, including tools to automate delivery of those services.</li>\n<li>Gain deep knowledge of our data platforms and tools to guide and enable stakeholders with their data needs.</li>\n<li>Work across our tech stack, which includes Kubernetes, Trino, Iceberg, Clickhouse, and PostgreSQL, with software built using Go, Javascript/Typescript, Python, and others.</li>\n<li>Collaborate with peers to reinforce a culture of exceptional delivery and accountability on the team.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>3-5+ years of experience as a software engineer with a focus on building and maintaining data infrastructure.</li>\n<li>Experience participating in technical initiatives in a cross-functional context, working with stakeholders to deliver value.</li>\n<li>Practical experience with data infrastructure components, such as Trino, Spark, Iceberg/Delta Lake, Kafka, Clickhouse, or PostgreSQL.</li>\n<li>Hands-on experience building and debugging data pipelines.</li>\n<li>Proficient using backend languages like Go, Python, or Typescript, along with strong SQL skills.</li>\n<li>Strong analytical skills, with a focus on understanding how data is used to drive business value.</li>\n<li>Solid communication skills, with the ability to explain technical concepts to both technical and non-technical audiences.</li>\n</ul>\n<p>Desirable Skills</p>\n<ul>\n<li>Experience with data orchestration and infrastructure platforms like Airflow and DBT.</li>\n<li>Experience deploying and managing services in Kubernetes.</li>\n<li>Familiarity with data governance processes, privacy requirements, or auditability.</li>\n<li>Interest in or knowledge of machine learning models and MLOps.</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law. We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals with disabilities. Please tell us if you require a reasonable accommodation to apply for a job. Examples of reasonable accommodations include, but are not limited to, changing the application process, providing documents in an alternate format, using a sign language interpreter, or using specialized equipment. If you require a reasonable accommodation to apply for a job, please contact us via e-mail at hr@cloudflare.com or via mail at 101 Townsend St. San Francisco, CA 94107.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_059293a1-afa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7527453","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data infrastructure","data pipelines","data products","Kubernetes","Trino","Iceberg","Clickhouse","PostgreSQL","Go","Javascript/Typescript","Python","SQL"],"x-skills-preferred":["data orchestration","infrastructure platforms","Airflow","DBT","machine learning models","MLOps"],"datePosted":"2026-04-18T15:50:12.541Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data infrastructure, data pipelines, data products, Kubernetes, Trino, Iceberg, Clickhouse, PostgreSQL, Go, Javascript/Typescript, Python, SQL, data orchestration, infrastructure platforms, Airflow, DBT, machine learning models, MLOps"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c571d7f7-d82"},"title":"Engineering Manager - Storage","description":"<p>At Databricks, we are building and running the world&#39;s best data and AI infrastructure platform. As an Engineering Manager, you will work with your team to build mission-critical Lakebase services on the Databricks Platform at scale.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Drive continuous delivery within a team of experts in storage technology, distributed systems and Rust.</li>\n<li>Manage the development and rollout of storage services that host millions of customer databases across dozens of regions</li>\n<li>Partner with peer engineering teams across Databricks to co-evolve Lakebase services with our global infrastructure.</li>\n<li>Lead operational excellence in 24/7 operation of our system</li>\n</ul>\n<p>The impact you will have:</p>\n<ul>\n<li>Hire great engineers to build an outstanding team.</li>\n<li>Support engineers in their career development by providing clear feedback and develop engineering leaders.</li>\n<li>Ensure high technical standards by instituting processes (architecture reviews, testing) and culture (engineering excellence).</li>\n<li>Work with engineering and product leadership to build a long-term roadmap.</li>\n<li>Coordinate execution and collaborate across teams to unblock cross-cutting projects.</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>Experience with building and shipping storage systems where correctness and performance are essential</li>\n<li>BS (or higher) in Computer Science, or a related field</li>\n<li>2+ years of experience building and leading a team of engineers working in a related system</li>\n<li>Experience with build, release and deployment infrastructure technologies such as Spinnaker, Jenkins, Airflow, Docker, Kubernetes, Terraform, Bazel, etc.</li>\n<li>Ability to attract, hire, and coach engineers who meet the Databricks hiring standards</li>\n<li>Comfort working on cross-functional projects with the ability to deeply understand product and customer personas</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c571d7f7-d82","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com/","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8476581002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["storage technology","distributed systems","Rust","Spinnaker","Jenkins","Airflow","Docker","Kubernetes","Terraform","Bazel"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:50.298Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London, United Kingdom"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"storage technology, distributed systems, Rust, Spinnaker, Jenkins, Airflow, Docker, Kubernetes, Terraform, Bazel"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f32fed2e-9ba"},"title":"Engineering Manager, Data Transformation","description":"<p>As an Engineering Manager of the Data Transformation team, you will lead a team of engineers, collaborate with infrastructure and product engineering orgs, and advance the Data Transformation roadmap and adoption at Stripe.</p>\n<p>You will be driving critical workstreams for Stripe&#39;s topmost priorities around delivering high quality, materialized datasets Stripe products and AI agents.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Delivering infrastructure and services that scale to our users&#39; needs with an eye on reliability and efficiency</li>\n<li>Leading and managing a team of talented engineers on the team, providing mentorship, guidance, and support to ensure their success</li>\n<li>Working with high-visibility teams and their stakeholders to support the Infrastructure&#39;s key engineering initiatives</li>\n<li>Understanding user needs and pain points to prioritize engineering work and deliver high quality solutions that meet user needs</li>\n<li>Driving the execution of projects, overseeing the entire development lifecycle from planning to delivery, while maintaining high standards of quality and timely completion</li>\n</ul>\n<p>You will also provide hands-on technical leadership (architecture/design, vision/direction/requirements setting, and incident response processes) for your reports, work with leaders across the company to create and drive toward the longer term vision of Stripe&#39;s Data Transformation roadmap, and foster a collaborative and inclusive work environment, promoting innovation, knowledge sharing, and continuous improvement within the team.</p>\n<p>We&#39;re looking for someone who has 1-3 years of experience managing teams that shipped and operated data pipelines and critical distributed system infrastructure, successfully recruited and built great teams, and works effectively cross-functionally and is able to think rigorously, communicate effectively, and make or coordinate hard decisions and trade-offs.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f32fed2e-9ba","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com/","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7688358","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kafka","Flink","Spark","Airflow","Python","SQL","API design"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:22.563Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"N/A"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kafka, Flink, Spark, Airflow, Python, SQL, API design"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_760c3e88-e35"},"title":"Senior Product Manager, Data","description":"<p>Job Title: Senior Product Manager, Data</p>\n<p>We are seeking a Senior Product Manager to support the development of CoreWeave&#39;s Enterprise Data Platform within the CIO organization. This role will contribute to building a scalable, high-performance data lake and data architecture, integrating data from key sources across Operations, Engineering, Sales, Finance, and other IT partners.</p>\n<p>As a Senior Product Manager for Data Infrastructure and Analytics, you will help drive data ingestion, transformation, governance, and analytics enablement. You will collaborate with engineering, analytics, finance, and business teams to help deliver data lake and pipeline orchestration solutions, ensuring accessible data for business insights.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Own and evangelize Data Platform and Business Analytics roadmap and strategy across CoreWeave</li>\n<li>Assist with the execution of CoreWeave&#39;s enterprise data architecture, helping enable the data lake and domain-driven data layer</li>\n<li>Support the development and enhancement of data ingestion, transformation, and orchestration pipelines for scalability, efficiency, and reliability</li>\n<li>Work with the Engineering and Data teams to maintain and enhance data pipelines for both structured and unstructured data, enabling efficient data movement across the organization</li>\n<li>Collaborate with Finance, GTM, Infrastructure, Data Center, and Supply Chain teams to help unify and model data from core systems (ERP, CRM, Asset Mgmt, Supply Chain systems, etc.)</li>\n<li>Contribute to data governance and quality initiatives, focusing on data consistency, lineage tracking, and compliance with security standards</li>\n<li>Support the BI and analytics layer by partnering with stakeholders to enable data products, dashboards, and reporting capabilities</li>\n<li>Help prioritize data-driven initiatives, ensuring alignment with business goals and operational needs in coordination with leadership</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>5+ years of experience in data product management, data architecture, or enterprise data engineering roles</li>\n<li>Familiarity with data lakes, data warehouses, ETL/ELT and streaming pipelines, and data governance frameworks</li>\n<li>Hands-on experience with modern data stack technologies (such as Snowflake, BigQuery, Databricks, Apache Spark, Airflow, DBT, Kafka)</li>\n<li>Understanding of data modeling, domain-driven design, and creating scalable data platforms</li>\n<li>Experience supporting the end-to-end data product lifecycle, including requirements gathering and implementation</li>\n<li>Strong collaboration skills with engineering, analytics, and business teams to help deliver data initiatives</li>\n<li>Awareness of data security, compliance, and governance best practices</li>\n<li>Understanding of BI and analytics platforms (such as Tableau, Looker, Power BI) and supporting self-service analytics</li>\n</ul>\n<p>Why CoreWeave?</p>\n<p>At CoreWeave, we work hard, have fun, and move fast! We&#39;re in an exciting stage of hyper-growth that you will not want to miss out on. We&#39;re not afraid of a little chaos, and we&#39;re constantly learning. Our team cares deeply about how we build our product and how we work together, which is represented through our core values:</p>\n<ul>\n<li>Be Curious at Your Core</li>\n<li>Act Like an Owner</li>\n<li>Empower Employees</li>\n<li>Deliver Best-in-Class Client Experiences</li>\n<li>Achieve More Together</li>\n</ul>\n<p>We support and encourage an entrepreneurial outlook and independent thinking. We foster an environment that encourages collaboration and provides the opportunity to develop innovative solutions to complex problems. As we get set for take off, the growth opportunities within the organization are constantly expanding. You will be surrounded by some of the best talent in the industry, who will want to learn from you, too. Come join us!</p>\n<p>Salary Range: $143,000 to $210,000</p>\n<p>Benefits:</p>\n<ul>\n<li>Medical, dental, and vision insurance - 100% paid for by CoreWeave</li>\n<li>Company-paid Life Insurance</li>\n<li>Voluntary supplemental life insurance</li>\n<li>Short and long-term disability insurance</li>\n<li>Flexible Spending Account</li>\n<li>Health Savings Account</li>\n<li>Tuition Reimbursement</li>\n<li>Ability to Participate in Employee Stock Purchase Program (ESPP)</li>\n<li>Mental Wellness Benefits through Spring Health</li>\n<li>Family-Forming support provided by Carrot</li>\n<li>Paid Parental Leave</li>\n<li>Flexible, full-service childcare support with Kinside</li>\n<li>401(k) with a generous employer match</li>\n<li>Flexible PTO</li>\n<li>Catered lunch each day in our office and data center locations</li>\n<li>A casual work environment</li>\n<li>A work culture focused on innovative disruption</li>\n</ul>\n<p>Workplace:</p>\n<p>While we prioritize a hybrid work environment, remote work may be considered for candidates located more than 30 miles from an office, based on role requirements for specialized skill sets. New hires will be invited to attend onboarding at one of our hubs within their first month. Teams also gather quarterly to support collaboration.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_760c3e88-e35","directApply":true,"hiringOrganization":{"@type":"Organization","name":"CoreWeave","sameAs":"https://www.coreweave.com","logo":"https://logos.yubhub.co/coreweave.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/coreweave/jobs/4649824006","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$143,000 to $210,000","x-skills-required":["data product management","data architecture","enterprise data engineering","data lakes","data warehouses","ETL/ELT and streaming pipelines","data governance frameworks","modern data stack technologies","Snowflake","BigQuery","Databricks","Apache Spark","Airflow","DBT","Kafka","data modeling","domain-driven design","scalable data platforms","BI and analytics platforms","Tableau","Looker","Power BI"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:48:58.405Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Livingston, NJ / New York, NY / Sunnyvale, CA / Bellevue, WA/San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data product management, data architecture, enterprise data engineering, data lakes, data warehouses, ETL/ELT and streaming pipelines, data governance frameworks, modern data stack technologies, Snowflake, BigQuery, Databricks, Apache Spark, Airflow, DBT, Kafka, data modeling, domain-driven design, scalable data platforms, BI and analytics platforms, Tableau, Looker, Power BI","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":143000,"maxValue":210000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6e6544bc-9bc"},"title":"Staff Machine Learning Engineer, Listings and Host Tools Data and AI","description":"<p>We&#39;re looking for a Staff Machine Learning Engineer to join our Listings and Host Tools Data and AI team. As a member of this team, you will support host personalization products and provide data-driven solutions to achieve a superior host experience on Airbnb.</p>\n<p>The Listings and Host Tools Data and AI team owns data pipelines and ML models and builds services for serving that are used in the above areas. We leverage open source, third-party, and homegrown ML models to improve the Host and Guest experience.</p>\n<p>As an ML engineer, you will partner closely with our data science, product partners, and other ML + data engineers on the team to execute on these opportunities in order to improve the Host and Guest product experience on Airbnb.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Working with large-scale structured and unstructured data to build and continuously improve cutting-edge Machine Learning models for Airbnb product, business, and operational use cases.</li>\n</ul>\n<ul>\n<li>Collaborating with cross-functional partners, including software engineers, product managers, operations, and data scientists, to identify opportunities for business impact, understand, refine, and prioritize requirements for machine learning models, drive engineering decisions, and quantify impact.</li>\n</ul>\n<ul>\n<li>Prototyping machine learning use cases for use in the product and working with stakeholders to iterate on requirements.</li>\n</ul>\n<ul>\n<li>Developing, productionizing, and operating Machine Learning models and pipelines at scale, including both batch and real-time use cases.</li>\n</ul>\n<ul>\n<li>Designing and building services and APIs to enable serving ML model-driven data to product use cases.</li>\n</ul>\n<p>We&#39;re looking for someone with 8+ years of industry experience in applied Machine Learning, including a Master&#39;s or Ph.D. in a relevant field. You should have experience in both Natural Language Processing and Computer Vision, as well as strong programming and data engineering skills.</p>\n<p>You should also have a deep understanding of Machine Learning best practices, algorithms, and domains, as well as experience with technologies such as TensorFlow, PyTorch, Kubernetes, Spark, Airflow, and data warehouses.</p>\n<p>If you&#39;re passionate about building end-to-end Machine Learning infrastructure and productionizing Machine Learning models, we&#39;d love to hear from you!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6e6544bc-9bc","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7454348","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$204,000-$255,000 USD","x-skills-required":["Machine Learning","Natural Language Processing","Computer Vision","Programming","Data Engineering","TensorFlow","PyTorch","Kubernetes","Spark","Airflow","Data Warehouses"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:48:25.034Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, Natural Language Processing, Computer Vision, Programming, Data Engineering, TensorFlow, PyTorch, Kubernetes, Spark, Airflow, Data Warehouses","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":204000,"maxValue":255000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5a5a8459-f04"},"title":"Engineering Manager of Managers, Data Platform","description":"<p>Job Description:</p>\n<p><strong>Who we are</strong></p>\n<p>Stripe is a financial infrastructure platform for businesses. Millions of companies - from the world’s largest enterprises to the most ambitious startups - use Stripe to accept payments, grow their revenue, and accelerate new business opportunities.</p>\n<p><strong>About the team</strong></p>\n<p>The Big Data Infrastructure organization is a globally distributed team of approximately 40 engineers spread across Dublin, Bangalore, Seattle, and San Francisco. This team is the backbone of the company’s data ecosystem, responsible for building, scaling, and maintaining the highly reliable platforms that power data storage, orchestration, and processing at scale.</p>\n<p>As the Head of Big Data Infra, you will lead a global, ~40-person engineering organization responsible for the foundational data platforms that drive the business. Reporting directly to the Head of Compute, you will define the strategic vision and roadmap for the company&#39;s data lake, orchestration pipelines, and batch computing environments.</p>\n<p>The team&#39;s technical portfolio spans four core domains:</p>\n<ul>\n<li>Datalake (Storage): Managing scalable cloud storage and metadata layers, leveraging Amazon S3, Apache Iceberg (metastore and integrations), SAL, and Hive Metastore (HMS).</li>\n</ul>\n<ul>\n<li>Data Orchestration: Ensuring robust pipeline execution and scheduling using Apache Airflow.</li>\n</ul>\n<ul>\n<li>Batch Compute Infra (Data Store): Maintaining foundational data infrastructure and legacy systems, including Hadoop.</li>\n</ul>\n<ul>\n<li>Batch Compute Experience (Data Processing): Optimizing and delivering powerful data processing environments utilizing Apache Spark and Apache Celeborn.</li>\n</ul>\n<p><strong>What you’ll do</strong></p>\n<p>You will move beyond day-to-day management to act as an industry leader, effectively advocating for your organization&#39;s mission and impact. You will be expected to see problems others don&#39;t and rally people to independently create solutions.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Set Strategic Vision: Define the scope, vision, and goals for your organization with little or no guidance. You will anticipate industry trends to influence Stripe&#39;s long-range plans and set direction on a multi-year timeframe.</li>\n</ul>\n<ul>\n<li>Lead at Scale: Manage the achievement of and accountability for broad swaths of programs. You will establish wide-ranging and scaled processes, anticipating and removing roadblocks across multiple teams.</li>\n</ul>\n<ul>\n<li>Drive Operational Excellence: Instill a culture of rigorous thinking and meticulous craftsmanship. You will ensure your organization drives constant improvement in team processes and maintains high standards of operational rigor.</li>\n</ul>\n<ul>\n<li>Indirect Influence: Use indirect influence to steer other teams toward making the right decisions for Stripe. You will effectively communicate your team&#39;s plan and how it links to Stripe&#39;s company vision to cross-functional stakeholders.</li>\n</ul>\n<ul>\n<li>Obsess Over Talent: Proactively invest in the development of the organization and its people at all levels. You will recruit world-class talent and coach your direct reports,who are themselves managers - to elevate the skills of the leadership team.</li>\n</ul>\n<ul>\n<li>Stewardship &amp; Culture: Act as an ambassador and advocate for Stripe, modeling ownership for all other Stripes. You will actively work to increase Stripe&#39;s inclusivity and diversity and use our operating principles to guide decision-making.</li>\n</ul>\n<p><strong>Who you are</strong></p>\n<p>We’re looking for someone who meets the minimum requirements to be considered for the role. If you meet these requirements, you are encouraged to apply. The preferred qualifications are a bonus, not a requirement.</p>\n<p><strong>Minimum requirements</strong></p>\n<ul>\n<li>Bachelor’s degree or equivalent practical experience with minimum 5 years of experience with software development.</li>\n</ul>\n<ul>\n<li>Minimum 5 years of experience in a technical leadership role; overseeing strategic projects.</li>\n</ul>\n<ul>\n<li>Minimum 3 years of Manager of Managers experience (managing other engineering managers).</li>\n</ul>\n<ul>\n<li>Experience building diverse teams to tackle challenging technical problems.</li>\n</ul>\n<ul>\n<li>Ability to thrive in a collaborative environment involving different stakeholders and subject matter experts.</li>\n</ul>\n<p><strong>Preferred qualifications</strong></p>\n<ul>\n<li>Strategic Ambiguity: Proven ability to translate chaos into clarity and navigate complex, high-impact work where you must define your own scope.</li>\n</ul>\n<ul>\n<li>Infrastructure at Scale: Successfully shipped and operated critical infrastructure with significant responsibility over funds or critical data.</li>\n</ul>\n<ul>\n<li>Cross-Functional Influence: A track record of getting other teams on board with your vision to support execution in a way that benefits the broader company.</li>\n</ul>\n<ul>\n<li>Curiosity: You enjoy learning and diving into the nuts-and-bolts of how things work (e.g., global money movement rails, currency conversion, or inter-company flows).</li>\n</ul>\n<ul>\n<li>Humility and Adaptability: You are humble and self-aware, with a history of adapting your management approach across different environments and seeking feedback to grow as a leader.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5a5a8459-f04","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7747391","x-work-arrangement":"onsite","x-experience-level":"executive","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Strategic vision","Technical leadership","Project management","Team management","Communication","Problem-solving","Infrastructure at scale","Cross-functional influence","Curiosity","Humility and adaptability"],"x-skills-preferred":["Apache Iceberg","Apache Airflow","Apache Spark","Apache Celeborn","Amazon S3","Hive Metastore","SAL","Cloud storage","Metadata layers","Data orchestration","Batch computing infrastructure","Legacy systems","Hadoop","Global money movement rails","Currency conversion","Inter-company flows"],"datePosted":"2026-04-18T15:47:47.234Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Strategic vision, Technical leadership, Project management, Team management, Communication, Problem-solving, Infrastructure at scale, Cross-functional influence, Curiosity, Humility and adaptability, Apache Iceberg, Apache Airflow, Apache Spark, Apache Celeborn, Amazon S3, Hive Metastore, SAL, Cloud storage, Metadata layers, Data orchestration, Batch computing infrastructure, Legacy systems, Hadoop, Global money movement rails, Currency conversion, Inter-company flows"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4b4378c3-f92"},"title":"Principal Software Engineer","description":"<p>We&#39;re looking for a Principal Software Engineer to join our Advertising, Company Intelligence, and Intent team. As a key member of our engineering team, you&#39;ll design and implement the core systems that power our real-time marketing platform.</p>\n<p>Your responsibilities will include:</p>\n<ul>\n<li>Designing and building distributed systems that process, enrich, and respond to billions of behavioral events per day in real time</li>\n<li>Developing high-performance APIs and services that support advertising, identity, and intent features across the Marketing Platform</li>\n<li>Leveraging machine learning and large language models (LLMs) to analyze behavioral data, classify content, extract signals, and enable intelligent decision-making</li>\n<li>Building intelligent agents using frameworks like LangGraph or MCP to reason over data and power user-facing insights</li>\n<li>Designing and operating data pipelines using tools like Kafka, Kinesis, and ClickHouse to support both streaming and batch workloads</li>\n<li>Driving quality, performance, scalability, and observability across all systems you own</li>\n<li>Collaborating cross-functionally with product managers, data scientists, and engineers to deliver customer-facing features and internal tooling</li>\n<li>Contributing to technical leadership and mentorship of teammates</li>\n</ul>\n<p>We&#39;re looking for someone with 8+ years of backend, data, or infrastructure engineering experience, or equivalent impact and leadership. You should have strong experience in at least one of the following areas:</p>\n<ul>\n<li>Distributed systems engineering</li>\n<li>Big data infrastructure</li>\n<li>Applied AI/ML</li>\n</ul>\n<p>You should also be proficient in one or more core languages (Java, Go, Python), have a solid grasp of SQL and large-scale data modeling, and familiarity with databases and tools such as ClickHouse, DynamoDB, Bigtable, Memcached, Kafka, Kinesis, Firehose, Airflow, Snowflake.</p>\n<p>Bonus points if you have experience in ad tech, real-time bidding (RTB), or programmatic systems, background in identity resolution, attribution, or behavioral analytics at scale, contributions to open source in ML, infrastructure, or data tooling, or strong product instincts and a passion for building tools that drive meaningful outcomes.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4b4378c3-f92","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8340521002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$163,800-$257,400 USD","x-skills-required":["Distributed systems engineering","Big data infrastructure","Applied AI/ML","Java","Go","Python","SQL","ClickHouse","DynamoDB","Bigtable","Memcached","Kafka","Kinesis","Firehose","Airflow","Snowflake"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:47:17.745Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bethesda, Maryland, United States; Remote US - PST; Waltham, Massachusetts, United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Distributed systems engineering, Big data infrastructure, Applied AI/ML, Java, Go, Python, SQL, ClickHouse, DynamoDB, Bigtable, Memcached, Kafka, Kinesis, Firehose, Airflow, Snowflake","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":163800,"maxValue":257400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f06742a2-a51"},"title":"Senior Software Engineer (Data Platform)","description":"<p>At Databricks, we are building and running the world&#39;s best data and AI infrastructure platform. Our engineering teams build technical products that fulfill real, important needs in the world. We develop and operate one of the largest scale software platforms. The fleet consists of millions of virtual machines, generating terabytes of logs and processing exabytes of data per day.</p>\n<p>As a Senior Software Engineer working on the Data Platform team, you will help build the Data Intelligence Platform for Databricks that will allow us to automate decision-making across the entire company. You will achieve this in collaboration with Databricks Product Teams, Data Science, Applied AI and many more. You will develop a variety of tools spanning logging, orchestration, data transformation, metric store, governance platforms, data consumption layers etc. You will do this using the latest, bleeding-edge Databricks product and other tools in the data ecosystem - the team also functions as a large, production, in-house customer that dog foods Databricks and guides the future direction of the product.</p>\n<p>The impact you will have:</p>\n<ul>\n<li>Design and run the Databricks metrics store that enables all business units and engineering teams to bring their detailed metrics into a common platform for sharing and aggregation, with high quality, introspection ability and query performance.</li>\n</ul>\n<ul>\n<li>Design and run the cross-company Data Intelligence Platform, which contains every business and product metric used to run Databricks. You’ll play a key role in developing the right balance of data protections and ease of shareability for the Data Intelligence Platform as we transition to a public company.</li>\n</ul>\n<ul>\n<li>Develop tooling and infrastructure to efficiently manage and run Databricks on Databricks at scale, across multiple clouds, geographies and deployment types. This includes CI/CD processes, test frameworks for pipelines and data quality, and infrastructure-as-code tooling.</li>\n</ul>\n<ul>\n<li>Design the base ETL framework used by all pipelines developed at the company.</li>\n</ul>\n<ul>\n<li>Partner with our engineering teams to provide leadership in developing the long-term vision and requirements for the Databricks product.</li>\n</ul>\n<ul>\n<li>Build reliable data pipelines and solve data problems using Databricks, our partner’s products and other OSS tools. Provide early feedback on the design and operations of these products.</li>\n</ul>\n<ul>\n<li>Establish conventions and create new APIs for telemetry, debug, feature and audit event log data, and evolve them as the product and underlying services change.</li>\n</ul>\n<ul>\n<li>Represent Databricks at academic and industrial conferences &amp; events.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f06742a2-a51","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/7647369002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["ETL frameworks","metrics stores","infrastructure management","data security","large-scale messaging systems","workflow or orchestration frameworks","Airflow","DBT","Kafka","RabbitMQ"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:46.775Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"ETL frameworks, metrics stores, infrastructure management, data security, large-scale messaging systems, workflow or orchestration frameworks, Airflow, DBT, Kafka, RabbitMQ"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7a6f0739-e83"},"title":"Senior Staff Machine Learning Engineer, Growth Platform Engineering","description":"<p>The Growth Platform team&#39;s vision is to drive long-term sustainable growth for the Airbnb community. Our mission is to build a best-in-class agentic system, and capabilities to support the growth of all Airbnb products, current and future.</p>\n<p>We achieve this by delivering highly personalised and relevant content and product experiences to the Airbnb community, both on and off of the Airbnb platform. The north star is full autonomy , where AI identifies opportunities, creates campaigns, personalises experiences, and optimises outcomes with minimal human intervention.</p>\n<p>As a machine learning engineer or scientist, your expertise will be pivotal in developing AI-powered solutions to shape the future of the Airbnb agentic growth platform with cutting-edge AI techniques. You will drive and guide the rest of the engineers to brainstorm, design and develop AI products and features from inception to production.</p>\n<p>Some example projects you will work on:</p>\n<ul>\n<li>AI-Powered Content Generation</li>\n<li>ML/AI Orchestration for Decisioning</li>\n<li>Proactive Marketing Analyst Agent</li>\n</ul>\n<p>A typical day will involve working with large scale structured and unstructured data; exploring, experimenting, building and continuously improving Machine Learning models and pipelines for Airbnb product, business and operational use cases.</p>\n<p>You will work collaboratively with cross-functional partners including product managers, operations and data scientists, to identify opportunities for business impact; understand, refine, and prioritise requirements for machine learning, and drive engineering decisions.</p>\n<p>Hands-on develop, productionise, and operate ML/AI models and pipelines at scale, including both batch and real-time use cases.</p>\n<p>Leverage third-party and in-house Machine Learning tools &amp; infrastructure to develop reusable, highly differentiating and high-performing Machine Learning systems, enable fast model development, low-latency serving and ease of model quality upkeep.</p>\n<p>Collaborate actively with engineers to apply ML / AI in their solutions to help validate ideas and guide to the right outcomes.</p>\n<p>Partner with ML/AI Engineers in foundations engineering to mentor and develop initiatives that make ML/AI applications a core discipline for non-ML/AI engineers.</p>\n<p>Your expertise will be crucial in developing AI-powered solutions to shape the future of the Airbnb agentic growth platform with cutting-edge AI techniques.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7a6f0739-e83","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7747259","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Machine Learning","AI","Python","Java","C++","TensorFlow","PyTorch","Kubernetes","Airflow","Kafka"],"x-skills-preferred":["Agentic and Automation","Agile Practice for AI Production","Infrastructure Acumen"],"datePosted":"2026-04-18T15:46:32.400Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - USA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Machine Learning, AI, Python, Java, C++, TensorFlow, PyTorch, Kubernetes, Airflow, Kafka, Agentic and Automation, Agile Practice for AI Production, Infrastructure Acumen"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_25f010f0-7d1"},"title":"Data Engineer","description":"<p>Why join us</p>\n<p>Brex is the intelligent finance platform that enables companies to spend smarter and move faster in more than 200 markets. By combining global corporate cards and banking with intuitive spend management, bill pay, and travel software, Brex enables founders and finance teams to accelerate operations, gain real-time visibility, and control spend effortlessly.</p>\n<p>Brex’s AI-native automation and world-class service eliminate manual expense and accounting tasks for customers so they can focus on what matters most. Tens of thousands of the world&#39;s best companies run on Brex, including DoorDash, Coinbase, Robinhood, Zoom, Plaid, Reddit, and SeatGeek.</p>\n<p>Working at Brex allows you to push your limits, challenge the status quo, and collaborate with some of the brightest minds in the industry. We’re committed to building a diverse team and inclusive culture and believe your potential should only be limited by how big you can dream. We make this a reality by empowering you with the tools, resources, and support you need to grow your career.</p>\n<p>Data at Brex</p>\n<p>Our Scientists and Engineers work together to make data , and insights derived from data , a core asset across Brex. But it&#39;s more than just crunching numbers. The Data team at Brex develops infrastructure, statistical models, and products using data. Our work is ingrained in Brex&#39;s decision-making process, the efficiency of our operations, our risk management policies, and the unparalleled experience we provide our customers.</p>\n<p>What You’ll Do</p>\n<p>As a Data Engineer at Brex, you will be a core contributor in transforming raw data into actionable insights for various departments across the organization. You&#39;ll collaborate closely with Data Scientists, Software Engineers, and business units to create efficient data models, pipelines, and analytics frameworks that drive the business forward. You also play a leading role in the design, implementation, and maintenance of Core Data tables, our high-quality, curated data source for a wide range of analytic applications.</p>\n<p>Where you’ll work</p>\n<p>This role will be based in our San Francisco office. We are a hybrid environment that combines the energy and connections of being in the office with the benefits and flexibility of working from home. We currently require a minimum of two coordinated days in the office per week, Wednesday and Thursday. Starting February 2, 2026, we will require three days per week in office - Monday, Wednesday and Thursday. As a perk, we also have up to four weeks per year of fully remote work!</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and maintain data models and pipelines that scale with the growing number of services, products, and changes in the company.</li>\n</ul>\n<ul>\n<li>Collaborate closely with Data Scientists, Data Analysts, and Business teams to understand their data needs, translating them into robust, efficient, scalable data solutions that enable ease of predictive analytics, data analysis, and metrics formulation.</li>\n</ul>\n<ul>\n<li>Maintain data documentation and definitions, building and ensuring that source-of-truth tables remain high quality for data science and reporting applications.</li>\n</ul>\n<ul>\n<li>Develop and enable integration with various data sources, allowing for more data-driven initiatives across the company.</li>\n</ul>\n<ul>\n<li>Apply best practices in data management to ensure the reliability and robustness of data utilized across various analytics applications.</li>\n</ul>\n<ul>\n<li>Set and proliferate company-wide standards for data relating to structure, quality, and expectations.</li>\n</ul>\n<ul>\n<li>Act as a liaison between the technical and non-technical teams, bridging gaps and ensuring that data solutions align with business objectives.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>3+ years of experience in Data Engineering, Data Analytics, or a related field such as Analytics Engineering.</li>\n</ul>\n<ul>\n<li>2+ years of experience working with modern data transformation tools like DBT.</li>\n</ul>\n<ul>\n<li>Advanced knowledge of databases and SQL with the ability to efficiently stage, process, and transform data.</li>\n</ul>\n<ul>\n<li>Experience integrating and orchestrating data workflows with various modern data tools and systems.</li>\n</ul>\n<ul>\n<li>Experience with data modeling, ETL/ELT processes, and data warehousing solutions.</li>\n</ul>\n<ul>\n<li>Experience working with a data warehouse such as Snowflake.</li>\n</ul>\n<ul>\n<li>Experience with a data workflow orchestrator tool such as Airflow.</li>\n</ul>\n<ul>\n<li>Experience with a programming language such as Python.</li>\n</ul>\n<ul>\n<li>Familiarity with BI tools such as Looker, Tableau, or similar platforms is a plus.</li>\n</ul>\n<ul>\n<li>Exceptional quantitative and analytical skills.</li>\n</ul>\n<ul>\n<li>Strong communication skills and ability to collaborate with various stakeholders, both technical and non-technical.</li>\n</ul>\n<p>Compensation:</p>\n<p>The expected salary range for this role is $120,800 - $151,000. However, the starting base pay will depend on a number of factors including the candidate’s location, skills, experience, market demands, and internal pay parity. Depending on the position offered, equity and other forms of compensation may be provided as part of a total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_25f010f0-7d1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Brex","sameAs":"https://brex.com/","logo":"https://logos.yubhub.co/brex.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/brex/jobs/8366850002","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$120,800 - $151,000","x-skills-required":["DBT","databases","SQL","data modeling","ETL/ELT processes","data warehousing solutions","Snowflake","Airflow","Python","BI tools","Looker","Tableau"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:18.514Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"DBT, databases, SQL, data modeling, ETL/ELT processes, data warehousing solutions, Snowflake, Airflow, Python, BI tools, Looker, Tableau","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":120800,"maxValue":151000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1d204fa1-067"},"title":"Data Engineer","description":"<p>Why join us</p>\n<p>Brex is the intelligent finance platform that enables companies to spend smarter and move faster in more than 200 markets. By combining global corporate cards and banking with intuitive spend management, bill pay, and travel software, Brex enables founders and finance teams to accelerate operations, gain real-time visibility, and control spend effortlessly.</p>\n<p>Data at Brex</p>\n<p>Our Scientists and Engineers work together to make data , and insights derived from data , a core asset across Brex. But it&#39;s more than just crunching numbers. The Data team at Brex develops infrastructure, statistical models, and products using data. Our work is ingrained in Brex&#39;s decision-making process, the efficiency of our operations, our risk management policies, and the unparalleled experience we provide our customers.</p>\n<p>What You’ll Do</p>\n<p>As a Data Engineer at Brex, you will be a core contributor in transforming raw data into actionable insights for various departments across the organization. You&#39;ll collaborate closely with Data Scientists, Software Engineers, and business units to create efficient data models, pipelines, and analytics frameworks that drive the business forward. You also play a leading role in the design, implementation, and maintenance of Core Data tables, our high-quality, curated data source for a wide range of analytic applications.</p>\n<p>Where you’ll work</p>\n<p>This role will be based in our Seattle office. We are a hybrid environment that combines the energy and connections of being in the office with the benefits and flexibility of working from home. We currently require a minimum of two coordinated days in the office per week, Wednesday and Thursday. Starting February 2, 2026, we will require three days per week in office - Monday, Wednesday and Thursday. As a perk, we also have up to four weeks per year of fully remote work!</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, build, and maintain data models and pipelines that scale with the growing number of services, products, and changes in the company.</li>\n</ul>\n<ul>\n<li>Collaborate closely with Data Scientists, Data Analysts, and Business teams to understand their data needs, translating them into robust, efficient, scalable data solutions that enable ease of predictive analytics, data analysis, and metrics formulation.</li>\n</ul>\n<ul>\n<li>Maintain data documentation and definitions, building and ensuring that source-of-truth tables remain high quality for data science and reporting applications.</li>\n</ul>\n<ul>\n<li>Develop and enable integration with various data sources, allowing for more data-driven initiatives across the company.</li>\n</ul>\n<ul>\n<li>Apply best practices in data management to ensure the reliability and robustness of data utilized across various analytics applications.</li>\n</ul>\n<ul>\n<li>Set and proliferate company-wide standards for data relating to structure, quality, and expectations.</li>\n</ul>\n<ul>\n<li>Act as a liaison between the technical and non-technical teams, bridging gaps and ensuring that data solutions align with business objectives.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>3+ years of experience in Data Engineering, Data Analytics, or a related field such as Analytics Engineering.</li>\n</ul>\n<ul>\n<li>2+ years of experience working with modern data transformation tools like DBT.</li>\n</ul>\n<ul>\n<li>Advanced knowledge of databases and SQL with the ability to efficiently stage, process, and transform data.</li>\n</ul>\n<ul>\n<li>Experience integrating and orchestrating data workflows with various modern data tools and systems.</li>\n</ul>\n<ul>\n<li>Experience with data modeling, ETL/ELT processes, and data warehousing solutions.</li>\n</ul>\n<ul>\n<li>Experience working with a data warehouse such as Snowflake.</li>\n</ul>\n<ul>\n<li>Experience with a data workflow orchestrator tool such as Airflow.</li>\n</ul>\n<ul>\n<li>Experience with a programming language such as Python.</li>\n</ul>\n<ul>\n<li>Familiarity with BI tools such as Looker, Tableau, or similar platforms is a plus.</li>\n</ul>\n<ul>\n<li>Exceptional quantitative and analytical skills.</li>\n</ul>\n<ul>\n<li>Strong communication skills and ability to collaborate with various stakeholders, both technical and non-technical.</li>\n</ul>\n<p>Compensation:</p>\n<p>The expected salary range for this role is $120,800 - $151,000. However, the starting base pay will depend on a number of factors including the candidate’s location, skills, experience, market demands, and internal pay parity. Depending on the position offered, equity and other forms of compensation may be provided as part of a total compensation package.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1d204fa1-067","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Brex","sameAs":"https://brex.com/","logo":"https://logos.yubhub.co/brex.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/brex/jobs/8510493002","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$120,800 - $151,000","x-skills-required":["DBT","databases","SQL","data modeling","ETL/ELT processes","data warehousing solutions","Snowflake","Airflow","Python","BI tools","Looker","Tableau"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:46:02.393Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Seattle, Washington, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"DBT, databases, SQL, data modeling, ETL/ELT processes, data warehousing solutions, Snowflake, Airflow, Python, BI tools, Looker, Tableau","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":120800,"maxValue":151000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_fc38e24f-97e"},"title":"Senior Machine Learning Engineer","description":"<p>We&#39;re looking for a Senior Machine Learning Engineer to join our Ads Engineering team. As a key member of our team, you will design and build production ML systems that power core experiences across the platform, including personalized recommendations, search, and ranking systems, intelligent advertising systems, and large-scale machine learning pipelines.</p>\n<p>Our team is responsible for building systems that operate at internet scale and directly influence user experience, advertiser value, and business outcomes. You will work on high-impact systems that improve ranking, recommendations, search relevance, prediction, content/user understanding, and optimization systems.</p>\n<p>As a Senior Machine Learning Engineer, you will:</p>\n<ul>\n<li>Design, build, and deploy production-grade machine learning models and systems at scale</li>\n<li>Own the full ML lifecycle: from problem definition and feature engineering to training, evaluation, deployment, and monitoring</li>\n<li>Build scalable data and model pipelines with strong reliability, observability, and automated retraining</li>\n<li>Work with large-scale datasets to improve ranking, recommendations, search relevance, prediction, content/user understanding, and optimization systems</li>\n<li>Partner cross-functionally with Product, Data Science, Infrastructure, and Engineering teams to translate complex problems into ML solutions</li>\n<li>Improve system performance across latency, throughput, and model quality metrics</li>\n<li>Research and apply state-of-the-art machine learning and AI techniques, including deep learning, graph &amp; transformers based, and LLM evaluation/alignment</li>\n</ul>\n<p>Basic Qualifications:</p>\n<ul>\n<li>3-5+ years of experience building, deploying, and operating machine learning systems in production</li>\n<li>Strong programming skills in Python, Java, Go, or similar languages, with solid software engineering fundamentals</li>\n<li>ML Fundamentals: a strong grasp of algorithms, from classic statistical learning (XGBoost, Random Forests, regressions) to DL architectures (Transformers, CNNs, GNNs)</li>\n<li>Hands-on experience with modern ML frameworks (e.g., PyTorch, TensorFlow)</li>\n<li>Experience designing scalable ML pipelines, data processing systems, and model serving infrastructure</li>\n<li>Ability to work cross-functionally and translate ambiguous product or business problems into technical solutions</li>\n<li>Experience improving measurable metrics through applied machine learning</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with recommender systems, search/ranking systems, advertising/auction systems, large-scale representation learning, or multimodal embedding systems</li>\n<li>Familiarity with distributed systems and large-scale data processing (Spark, Kafka, Ray, Airflow, BigQuery, Redis, etc.)</li>\n<li>Experience working with real-time systems and low-latency production environments</li>\n<li>Background in feature engineering, model optimization, and production monitoring</li>\n<li>Experience with LLM/Gen AI techniques, including but not limited to LLM evaluation, alignment, fine-tuning, knowledge distillation, RAG/agentic systems and productionizing LLM-powered products at scale</li>\n<li>Advanced degree in Computer Science, Machine Learning, or related quantitative field</li>\n</ul>\n<p>Potential Teams:</p>\n<ul>\n<li>Ads Measurement Modeling</li>\n<li>Ads Targeting and Retrieval</li>\n<li>Advertiser Optimization</li>\n<li>Ads Marketplace Quality</li>\n<li>Ads Creative Effectiveness</li>\n<li>Ads Foundational Representations</li>\n<li>Ads Content Understanding</li>\n<li>Ads Ranking</li>\n<li>Feed Relevance</li>\n<li>Search and Answers Relevance</li>\n<li>ML Understanding</li>\n<li>Notifications Relevance</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Comprehensive Healthcare Benefits and Income Replacement Programs</li>\n<li>401k with Employer Match</li>\n<li>Global Benefit programs that fit your lifestyle, from workspace to professional development to caregiving support</li>\n<li>Family Planning Support</li>\n<li>Gender-Affirming Care</li>\n<li>Mental Health &amp; Coaching Benefits</li>\n<li>Flexible Vacation &amp; Paid Volunteer Time Off</li>\n<li>Generous Paid Parental Leave</li>\n</ul>\n<p>Pay Transparency:</p>\n<p>This job posting may span more than one career level. In addition to base salary, this job is eligible to receive equity in the form of restricted stock units, and depending on the position offered, it may also be eligible to receive a commission. Additionally, Reddit offers a wide range of benefits to U.S.-based employees, including medical, dental, and vision insurance, 401(k) program with employer match, generous time off for vacation, and parental leave. To learn more, please visit https://www.redditinc.com/careers/. To provide greater transparency to candidates, we share base salary ranges for all US-based job postings regardless of state. We set standard base pay ranges for all roles based on function, level, and country location, benchmarked against similar stage growth companies. Final offer amounts are determined by multiple factors including, skills, depth of work experience and relevant licenses/credentials, and may vary from the amounts listed below. The base salary range for this position is $216,700-$303,400 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_fc38e24f-97e","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/6960831","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$216,700-$303,400 USD","x-skills-required":["Python","Java","Go","PyTorch","TensorFlow","XGBoost","Random Forests","Regressions","Transformers","CNNs","GNNs","Spark","Kafka","Ray","Airflow","BigQuery","Redis"],"x-skills-preferred":["Recommender systems","Search/ranking systems","Advertising/auction systems","Large-scale representation learning","Multimodal embedding systems","Distributed systems","Large-scale data processing","Real-time systems","Low-latency production environments","Feature engineering","Model optimization","Production monitoring","LLM/Gen AI techniques","LLM evaluation","Alignment","Fine-tuning","Knowledge distillation","RAG/agentic systems"],"datePosted":"2026-04-18T15:45:58.533Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote - United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Go, PyTorch, TensorFlow, XGBoost, Random Forests, Regressions, Transformers, CNNs, GNNs, Spark, Kafka, Ray, Airflow, BigQuery, Redis, Recommender systems, Search/ranking systems, Advertising/auction systems, Large-scale representation learning, Multimodal embedding systems, Distributed systems, Large-scale data processing, Real-time systems, Low-latency production environments, Feature engineering, Model optimization, Production monitoring, LLM/Gen AI techniques, LLM evaluation, Alignment, Fine-tuning, Knowledge distillation, RAG/agentic systems","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":216700,"maxValue":303400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_17d99112-d46"},"title":"Software Engineer, Product Catalogs","description":"<p>We are looking for a skilled backend software engineer to join the Product Catalogs team at Reddit. Our team builds products and infrastructure that enable retail advertisers to succeed on Reddit.</p>\n<p>As a software engineer on this team, you will have the opportunity to work on projects such as catalog system scaling, catalog management, and product enhancement. You will develop, maintain, and scale our product catalogs backend, contribute to the development of features to make our product easier to use, and produce robust and sustainable code.</p>\n<p>To be successful in this role, you will need a bachelor&#39;s degree or equivalent experience in a quantitative or computer science-related field, 4+ years of full-time backend software engineering experience in a scalable computing environment, and strong communication and collaboration skills.</p>\n<p>We offer a dynamic work environment, opportunities for professional growth and development, a competitive salary and benefits package, and flexible work arrangements.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_17d99112-d46","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Reddit","sameAs":"https://www.redditinc.com","logo":"https://logos.yubhub.co/redditinc.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/reddit/jobs/7761320","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Scala","Go","gRPC","Thrift","Baseplate","Kafka","Postgres","BigQuery","Redis","TiDB","Kubernetes","Airflow"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:45:40.737Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Amsterdam, Netherlands"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Scala, Go, gRPC, Thrift, Baseplate, Kafka, Postgres, BigQuery, Redis, TiDB, Kubernetes, Airflow"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c53ecdd3-dc7"},"title":"Scale Solution Engineer","description":"<p>As a Scale Solution Engineer at Databricks, you will play a critical role in advising customers during their onboarding process. You will work directly with customers to help them onboard and deploy Databricks in their production environment.</p>\n<p>Your impact will be significant, ensuring new customers have an excellent experience by providing technical assistance early in their journey. You will become an expert on the Databricks Platform and guide customers in making the best technical decisions. You will also work directly with multiple customers concurrently to provide technical solutions.</p>\n<p>To succeed in this role, you will need:</p>\n<ul>\n<li>An undergraduate degree or higher in Computer Science, Information Systems, or relevant experience</li>\n<li>1+ years experience in a technical role, preferably in the data or cloud field</li>\n<li>Knowledge of at least one of the public cloud platforms AWS, Azure, or GCP</li>\n<li>Knowledge of a programming language such as Python, Scala, or SQL</li>\n<li>Knowledge of end-to-end data analytics workflow</li>\n<li>Hands-on professional or academic experience in one or more of the following: Data Engineering technologies (e.g., ETL, DBT, Spark, Airflow), Data Warehousing technologies (e.g., SQL, Stored Procedures, Redshift, Snowflake)</li>\n<li>Excellent time management and prioritization skills</li>\n<li>Excellent written and verbal communication</li>\n</ul>\n<p>Bonus: Knowledge of Data Science and Machine Learning (e.g., build and deploy ML Models)</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c53ecdd3-dc7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8408817002","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["public cloud platforms","AWS","Azure","GCP","Python","Scala","SQL","Data Engineering technologies","ETL","DBT","Spark","Airflow","Data Warehousing technologies","Stored Procedures","Redshift","Snowflake"],"x-skills-preferred":["Data Science","Machine Learning"],"datePosted":"2026-04-18T15:44:58.601Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Rica"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"public cloud platforms, AWS, Azure, GCP, Python, Scala, SQL, Data Engineering technologies, ETL, DBT, Spark, Airflow, Data Warehousing technologies, Stored Procedures, Redshift, Snowflake, Data Science, Machine Learning"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bd9625d9-99b"},"title":"ML Infrastructure Engineer, Safeguards","description":"<p>We are seeking a Machine Learning Infrastructure Engineer to join our Safeguards organization, where you&#39;ll build and scale the critical infrastructure that powers our AI safety systems.</p>\n<p>As part of the Safeguards team, you&#39;ll design and implement ML infrastructure that powers Claude safety. Your work will directly contribute to making AI systems more trustworthy and aligned with human values, ensuring our models operate safely as they become more capable.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build scalable ML infrastructure to support real-time and batch classifier and safety evaluations across our model ecosystem</li>\n<li>Build monitoring and observability tools to track model performance, data quality, and system health for safety-critical applications</li>\n<li>Collaborate with research teams to productionize safety research, translating experimental safety techniques into robust, scalable systems</li>\n<li>Optimize inference latency and throughput for real-time safety evaluations while maintaining high reliability standards</li>\n<li>Implement automated testing, deployment, and rollback systems for ML models in production safety applications</li>\n<li>Partner with Safeguards, Security, and Alignment teams to understand requirements and deliver infrastructure that meets safety and production needs</li>\n<li>Contribute to the development of internal tools and frameworks that accelerate safety research and deployment</li>\n</ul>\n<p>You may be a good fit if you:</p>\n<ul>\n<li>Have 5+ years of experience building production ML infrastructure, ideally in safety-critical domains like fraud detection, content moderation, or risk assessment</li>\n<li>Are proficient in Python and have experience with ML frameworks like PyTorch, TensorFlow, or JAX</li>\n<li>Have hands-on experience with cloud platforms (AWS, GCP) and container orchestration (Kubernetes)</li>\n<li>Understand distributed systems principles and have built systems that handle high-throughput, low-latency workloads</li>\n<li>Have experience with data engineering tools and building robust data pipelines (e.g., Spark, Airflow, streaming systems)</li>\n<li>Are results-oriented, with a bias towards reliability and impact in safety-critical systems</li>\n<li>Enjoy collaborating with researchers and translating cutting-edge research into production systems</li>\n<li>Care deeply about AI safety and the societal impacts of your work</li>\n</ul>\n<p>Strong candidates may have experience with:</p>\n<ul>\n<li>Working with large language models and modern transformer architectures</li>\n<li>Implementing A/B testing frameworks and experimentation infrastructure for ML systems</li>\n<li>Developing monitoring and alerting systems for ML model performance and data drift</li>\n<li>Building automated labeling systems and human-in-the-loop workflows</li>\n<li>Experience in trust &amp; safety, fraud prevention, or content moderation domains</li>\n<li>Knowledge of privacy-preserving ML techniques and compliance requirements</li>\n<li>Contributing to open-source ML infrastructure projects</li>\n</ul>\n<p>We encourage you to apply even if you do not believe you meet every single qualification. Not all strong candidates will meet every single qualification as listed.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bd9625d9-99b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com/","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4778843008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000-$405,000 USD","x-skills-required":["Python","PyTorch","TensorFlow","JAX","Cloud platforms (AWS, GCP)","Container orchestration (Kubernetes)","Distributed systems principles","Data engineering tools (Spark, Airflow, streaming systems)"],"x-skills-preferred":["Large language models and modern transformer architectures","A/B testing frameworks and experimentation infrastructure for ML systems","Monitoring and alerting systems for ML model performance and data drift","Automated labeling systems and human-in-the-loop workflows","Trust & safety, fraud prevention, or content moderation domains","Privacy-preserving ML techniques and compliance requirements","Open-source ML infrastructure projects"],"datePosted":"2026-04-18T15:44:06.907Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, PyTorch, TensorFlow, JAX, Cloud platforms (AWS, GCP), Container orchestration (Kubernetes), Distributed systems principles, Data engineering tools (Spark, Airflow, streaming systems), Large language models and modern transformer architectures, A/B testing frameworks and experimentation infrastructure for ML systems, Monitoring and alerting systems for ML model performance and data drift, Automated labeling systems and human-in-the-loop workflows, Trust & safety, fraud prevention, or content moderation domains, Privacy-preserving ML techniques and compliance requirements, Open-source ML infrastructure projects","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_93c1356c-a95"},"title":"Principal Software Engineer, Web Data - Tech Lead","description":"<p>We&#39;re looking for an exceptional Principal Software Engineer to serve as the de facto Technical Lead for our Web Data Acquisition (WDA) team. This is a highly visible, hands-on technical leadership role where you&#39;ll own the architectural direction for crawling systems, evolve and unify crawling platforms into a best-in-class stack, and elevate a high-performing engineering team.</p>\n<p>As a Principal Software Engineer, you&#39;ll solve complex distributed systems challenges, build modular tooling that accelerates delivery, and set the standard for observability and operational excellence. You&#39;ll have a dedicated manager handling all HR and administrative responsibilities. A product manager connects business needs with technical work. Your focus is 100% technical leadership, mentorship, and hands-on execution.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Technical Leadership &amp; System Design: Proven experience building web crawling or large-scale data systems from scratch. Strong architectural skills designing scalable, fault-tolerant distributed systems. Track record leading complex technical initiatives and driving architecture direction for teams.</li>\n</ul>\n<ul>\n<li>Data Engineering Expertise: Deep background in large-scale data engineering (terabytes daily). Hands-on experience with cloud data warehouses (BigQuery, Snowflake). Experience with Apache Kafka, Kubernetes (GKE/EKS), and orchestration tools (Airflow).</li>\n</ul>\n<ul>\n<li>Web Crawling &amp; Data Extraction: Deep expertise in web crawling technologies and advanced scraping (Scrapy or similar). Experience extracting structured/unstructured web data and SERP extraction. Knowledge of proxy infrastructure management, anti-bot detection, and ethical crawling.</li>\n</ul>\n<ul>\n<li>Leadership &amp; Team Development: Experience mentoring engineers at all levels and fostering collaborative culture. Strong ability to influence technical direction and establish best practices. Track record hiring, coaching, and developing senior engineers.</li>\n</ul>\n<p>Ideal Candidate Profile:</p>\n<ul>\n<li>10+ years software engineering experience. 5+ years focused on data engineering. 3+ years in senior/principal-level technical leadership.</li>\n</ul>\n<ul>\n<li>Strong CS fundamentals (algorithms, data structures, distributed systems). Self-starter who thrives in fast-paced environments.</li>\n</ul>\n<p>Core Technical Stack:</p>\n<ul>\n<li>Python &amp; Java</li>\n<li>Apache Kafka</li>\n<li>GCP (BigQuery, GKE, Vertex AI)</li>\n<li>Snowflake &amp; Starburst/Trino</li>\n<li>Terraform</li>\n<li>Scrapy / Web Scraping Frameworks</li>\n<li>Proxy Management Systems</li>\n<li>Distributed Systems &amp; Kubernetes</li>\n<li>Apache Airflow</li>\n<li>Large-Scale ETL Pipelines</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_93c1356c-a95","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8378092002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$163,800-$257,400 USD","x-skills-required":["Python","Java","Apache Kafka","Kubernetes","GCP","Snowflake","Terraform","Scrapy","Proxy Management Systems","Distributed Systems","Apache Airflow","Large-Scale ETL Pipelines"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:43:50.896Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Apache Kafka, Kubernetes, GCP, Snowflake, Terraform, Scrapy, Proxy Management Systems, Distributed Systems, Apache Airflow, Large-Scale ETL Pipelines","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":163800,"maxValue":257400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9ce3bb01-4a1"},"title":"Scale Solutions Engineer","description":"<p>At Databricks, we aim to empower our customers to solve the world&#39;s most challenging data problems using the Data Intelligence platform. As a Scale Solution Engineer, you will be critical in advising customers during their onboarding. You will work directly with customers to help them onboard and deploy Databricks in their production environment and accelerate Databricks features adoption.</p>\n<p>The impact you will have:</p>\n<ul>\n<li>Ensure new customers have an excellent experience by providing technical assistance early in their journey</li>\n<li>Become an expert on the Databricks Platform and guide customers in making the best technical decisions</li>\n<li>Work directly with multiple customers concurrently to provide technical solutions</li>\n</ul>\n<p>What we look for:</p>\n<ul>\n<li>Undergraduate degree or higher in Computer Science, Information Systems, or relevant experience</li>\n<li>3+ years experience in a customer-facing technical role in pre-sales, professional services, consulting or customer success</li>\n<li>Experience in one or more of the following:</li>\n</ul>\n<ul>\n<li>Solid understanding of the end-to-end data analytics workflow</li>\n<li>Excellent time management and prioritization skills</li>\n<li>Knowledge of public cloud platforms AWS, Azure or GCP would be a plus</li>\n<li>Knowledge of a programming language - Python, Scala, or SQL</li>\n<li>Knowledge of end-to-end data analytics workflow</li>\n<li>Hands-on professional or academic experience in one or more of the following:</li>\n</ul>\n<ul>\n<li>Data Engineering technologies (e.g., ETL, DBT, Spark, Airflow)</li>\n<li>Data Warehousing technologies (e.g., SQL, Stored Procedures, Redshift, Snowflake)</li>\n<li>Excellent written and verbal communication, in English and Portuguese</li>\n<li>Bonus - Knowledge of Data Science and Machine Learning (e.g., build and deploy ML Models).</li>\n<li>Databricks certification(s)</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9ce3bb01-4a1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8391865002","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Databricks","Data Engineering","Data Warehousing","Python","Scala","SQL","AWS","Azure","GCP","ETL","DBT","Spark","Airflow","Redshift","Snowflake","English","Portuguese"],"x-skills-preferred":["Data Science","Machine Learning"],"datePosted":"2026-04-18T15:43:14.531Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Sao Paulo, Brazil"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Databricks, Data Engineering, Data Warehousing, Python, Scala, SQL, AWS, Azure, GCP, ETL, DBT, Spark, Airflow, Redshift, Snowflake, English, Portuguese, Data Science, Machine Learning"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6a19ea2e-6ab"},"title":"Software Engineer, Data","description":"<p>As a Software Engineer, Data at Airtable, you&#39;ll make an enormous contribution to our data engineering efforts. You&#39;ll design and own mission-critical data pipelines to enable decision-making, partner with company leaders to create scalable data solutions, and launch innovative alerting and visualization solutions.</p>\n<p>You&#39;ll work across our engineering organisation and stakeholders from data science, growth, sales, marketing, and product to understand the data needs of the business and produce pipelines, data marts, and other solutions that enable better decision-making.</p>\n<p>You&#39;ll design and maintain our foundational business tables in order to simplify analysis and reporting across the entire company, including AI usage metrics surfaced to executive stakeholders.</p>\n<p>You&#39;ll use AI tools as a daily part of how you work, from LLM-assisted pipeline development and debugging to exploring our catalog through AI-powered discovery, and bring a curiosity for where this tooling is heading next.</p>\n<p>You&#39;ll build and enforce a pattern language across our data stack, ensuring pipelines and tables are consistent, accurate, and well-understood.</p>\n<p>You&#39;ll continue to improve the performance and reliability of our data warehouse.</p>\n<p>You&#39;ll partner with data scientists, analytics engineers, and business stakeholders to translate ambiguous business questions into well-scoped data solutions.</p>\n<p>You have 3-8+ years of professional experience designing, creating, and maintaining scalable data pipelines, preferably in Airflow.</p>\n<p>You&#39;ve wrangled enough data to understand how often the complex systems that produce it can go wrong, and you build with that in mind.</p>\n<p>You are proficient in at least one programming language (preferably Python) and are willing to pick up others as the work demands.</p>\n<p>You are highly effective with SQL and understand how to write and tune complex queries.</p>\n<p>You&#39;re genuinely curious about how AI is reshaping data engineering and you&#39;re actively experimenting, not just watching from the sidelines.</p>\n<p>Whether that&#39;s using LLMs to write and debug pipelines faster, thinking through how to model agent behaviour as data, or exploring what smarter data discovery could look like, you bring enthusiasm for figuring it out.</p>\n<p>You&#39;re passionate and thoughtful about building systems that enhance human understanding.</p>\n<p>You communicate with clarity and precision in written form and have experience conveying findings through graphs and visualisations.</p>\n<p>Compensation awarded to successful candidates will vary based on their work location, relevant skills, and experience. Our total compensation package also includes the opportunity to receive benefits, restricted stock units, and may include incentive compensation.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6a19ea2e-6ab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airtable","sameAs":"https://www.airtable.com/","logo":"https://logos.yubhub.co/airtable.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airtable/jobs/8124953002","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$162,000-$199,800 USD","x-skills-required":["Python","SQL","Airflow","LLMs","Data engineering","Data science","Growth","Sales","Marketing","Product"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:41:20.358Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA; Austin, TX; New York, NY"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Airflow, LLMs, Data engineering, Data science, Growth, Sales, Marketing, Product","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":162000,"maxValue":199800,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_cece3778-5b8"},"title":"Finance Systems Integration Engineer","description":"<p>We are seeking an experienced Finance Systems Integration Engineer to support our finance systems transformation at one of the fastest-growing AI companies. You&#39;ll design and build integrations connecting our ERP platform with critical financial applications and support our ERP implementation initiatives.</p>\n<p>As you master our integration landscape, you&#39;ll have opportunities to expand into Claude-powered AI automation and data pipeline development.</p>\n<p>You&#39;ll build the integration backbone for one of the fastest-growing AI companies, with a front-row seat to how Claude transforms financial operations. This is a foundational role where you&#39;ll shape our integration architecture from the ground up, then expand into cutting-edge AI automation as our needs evolve.</p>\n<p><strong>Responsibilities</strong></p>\n<p><strong>Core Focus: Integration Development &amp; ERP Support</strong></p>\n<ul>\n<li>Design, build, and maintain integrations connecting ERP systems with downstream applications including ZipHQ, Brex, Navan, Clearwater, Payroll systems, Salesforce, and other critical financial platforms using Workato, MuleSoft, or similar iPaaS solutions</li>\n</ul>\n<ul>\n<li>Support integration development and testing during the ERP implementation projects</li>\n</ul>\n<ul>\n<li>Develop and maintain REST APIs, webhooks, and OAuth 2.0 authentication flows for secure system-to-system communication</li>\n</ul>\n<ul>\n<li>Implement real-time and batch integration patterns supporting high-volume financial transactions</li>\n</ul>\n<ul>\n<li>Establish monitoring, alerting, and error-handling frameworks to ensure integration reliability and data integrity</li>\n</ul>\n<ul>\n<li>Document integration architectures, data flows, API specifications, and troubleshooting procedures</li>\n</ul>\n<ul>\n<li>Collaborate with implementation consulting partners and vendors on technical integration requirements</li>\n</ul>\n<p><strong>Additional Scope: AI Automation &amp; Data Infrastructure</strong></p>\n<ul>\n<li>Build and deploy Claude-powered AI agents that automate financial operations including intelligent document processing, workflow automation, financial audit and reconciliations, and self-service reporting</li>\n</ul>\n<ul>\n<li>Design agentic workflows that leverage Claude API capabilities integrated with ERP platform data and processes</li>\n</ul>\n<ul>\n<li>Create automated validation and quality assurance processes for AI-generated outputs</li>\n</ul>\n<ul>\n<li>Partner with Finance teams to identify automation opportunities and translate requirements into AI agent solutions</li>\n</ul>\n<ul>\n<li>Support data pipeline development using Airflow for workflow orchestration and dbt for data transformation</li>\n</ul>\n<ul>\n<li>Build and maintain data flows from ERP and other financial systems into BigQuery for analytics and reporting</li>\n</ul>\n<ul>\n<li>Implement data quality checks and testing frameworks for financial data pipelines</li>\n</ul>\n<ul>\n<li>Collaborate with Data Infrastructure team on pipeline architecture, performance optimization, and security monitoring</li>\n</ul>\n<ul>\n<li>Support executive dashboards and financial analytics by ensuring timely, accurate data delivery</li>\n</ul>\n<p><strong>Governance &amp; Collaboration</strong></p>\n<ul>\n<li>Maintain comprehensive documentation for integrations, AI agents, and data pipelines</li>\n</ul>\n<ul>\n<li>Support internal and external audits with technical evidence and system access reviews</li>\n</ul>\n<ul>\n<li>Collaborate with Finance Systems Engineers on operational support, troubleshooting, and enhancement requests</li>\n</ul>\n<ul>\n<li>Partner with Finance Operations, Accounting, FP&amp;A, Engineering, and Data Infrastructure teams to deliver holistic solutions</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>8+ years of experience in integration development, data engineering, or systems engineering roles</li>\n</ul>\n<ul>\n<li>Hands-on experience with iPaaS platforms such as Workato, MuleSoft, Dell Boomi, or similar integration tools</li>\n</ul>\n<ul>\n<li>Strong programming skills in Python and/or JavaScript/TypeScript for building custom integrations, APIs, and automation scripts</li>\n</ul>\n<ul>\n<li>Experience with data pipeline tools including Airflow for orchestration and dbt for transformation</li>\n</ul>\n<ul>\n<li>Working knowledge of cloud data platforms such as BigQuery, Snowflake, or Databricks</li>\n</ul>\n<ul>\n<li>Understanding of REST API design patterns, webhooks, OAuth 2.0, and modern integration architectures</li>\n</ul>\n<ul>\n<li>Familiarity with ERP systems (Oracle Fusion, Workday Financials, or similar) and financial business processes</li>\n</ul>\n<ul>\n<li>Strong problem-solving skills with ability to debug complex integration issues across multiple systems</li>\n</ul>\n<ul>\n<li>Excellent communication skills to collaborate with technical and business stakeholders</li>\n</ul>\n<p><strong>Preferred Qualifications</strong></p>\n<ul>\n<li>Experience with high-growth technology companies scaling through rapid revenue expansion (5x-10x growth)</li>\n</ul>\n<ul>\n<li>Background in AI/ML companies with familiarity in modern SaaS business models including consumption-based pricing, usage metering platforms, and marketplace billing</li>\n</ul>\n<ul>\n<li>Hands-on experience with specific platforms: Workday Financials (Workday Studio, EIB, custom reports, Prism Analytics)</li>\n</ul>\n<ul>\n<li>Technical expertise with modern finance tech stack including Stripe, Salesforce, Zuora RevPro, Zip Procurement, Clearwater treasury systems, Pigment planning tools, Numeric close management</li>\n</ul>\n<ul>\n<li>Programming skills in Python / JavaScript, or similar languages for building custom integrations, APIs, and automation scripts</li>\n</ul>\n<ul>\n<li>Experience with AI/LLM integration for financial operations, including document processing, data extraction, intelligent automation, and agentic workflows (familiarity with Claude models and API is a plus)</li>\n</ul>\n<ul>\n<li>Hands-on experience with modern data stack tools: BigQuery/Snowflake/Databricks, dbt for data transformation, Airflow for workflow orchestration</li>\n</ul>\n<ul>\n<li>Professional certifications such as Workato, Workday integrations, or relevant technical credentials</li>\n</ul>\n<ul>\n<li>Bachelor&#39;s or Master&#39;s degree in Computer Science, Information Systems, Accounting, Finance, Engineering, or related technical/business field</li>\n</ul>\n<ul>\n<li>Experience with business intelligence and financial reporting tools (Hex, Looker, Tableau, Power BI) for executive dashboards and financial analytics</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_cece3778-5b8","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://anthropic.com","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5155195008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$205,000-$265,000 USD","x-skills-required":["integration development","data engineering","systems engineering","iPaaS platforms","Python","JavaScript/TypeScript","Airflow","dbt","BigQuery","Snowflake","Databricks","REST API design patterns","webhooks","OAuth 2.0","modern integration architectures","ERP systems","financial business processes"],"x-skills-preferred":["high-growth technology companies","AI/ML companies","SaaS business models","consumption-based pricing","usage metering platforms","marketplace billing","Workday Financials","Stripe","Salesforce","Zuora RevPro","Zip Procurement","Clearwater treasury systems","Pigment planning tools","Numeric close management","Python/JavaScript","AI/LLM integration","document processing","data extraction","intelligent automation","agentic workflows","Claude models","API","BigQuery/Snowflake/Databricks","professional certifications","Workato","Workday integrations","technical credentials","Computer Science","Information Systems","Accounting","Finance","Engineering","business intelligence","financial reporting tools"],"datePosted":"2026-04-18T15:39:50.764Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"integration development, data engineering, systems engineering, iPaaS platforms, Python, JavaScript/TypeScript, Airflow, dbt, BigQuery, Snowflake, Databricks, REST API design patterns, webhooks, OAuth 2.0, modern integration architectures, ERP systems, financial business processes, high-growth technology companies, AI/ML companies, SaaS business models, consumption-based pricing, usage metering platforms, marketplace billing, Workday Financials, Stripe, Salesforce, Zuora RevPro, Zip Procurement, Clearwater treasury systems, Pigment planning tools, Numeric close management, Python/JavaScript, AI/LLM integration, document processing, data extraction, intelligent automation, agentic workflows, Claude models, API, BigQuery/Snowflake/Databricks, professional certifications, Workato, Workday integrations, technical credentials, Computer Science, Information Systems, Accounting, Finance, Engineering, business intelligence, financial reporting tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":205000,"maxValue":265000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a98d4ace-d27"},"title":"Senior Data Engineer","description":"<p>We are looking for a Senior Data Engineer to join our high-performing data enablement team. As a Senior Data Engineer, you will play a pivotal role within the Data team that powers Yuno and its payment platform, while helping co-design and implement an architecture that scales with the product and the company.</p>\n<p>The stack is modern: StarRocks as our primary analytical layer, Flink for processing, DBT for transformation, Airflow for orchestration and various tooling for surfacing insights.</p>\n<p>You&#39;ll be working on things that matter and are technically interesting:</p>\n<ul>\n<li><p>Design and build data pipelines for large volumes of payment data that are performant, reliable, and correct , not just fast.</p>\n</li>\n<li><p>Own end-to-end data flows: from ingestion and transformation through to the outputs that Finance, Product, and clients depend on.</p>\n</li>\n<li><p>Drive data quality across your domain with tooling.</p>\n</li>\n<li><p>Work cross-functionally with Product, Finance and enable other Engineering teams via a &#39;consulting&#39; style model.</p>\n</li>\n<li><p>Contribute to how the team works , code review culture, CI/CD standards, ADRs, how we handle incidents , we&#39;re building these practices now and senior engineers shape them.</p>\n</li>\n<li><p>Help onboard and level up engineers around you; there&#39;s real opportunity to make an impact here.</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a98d4ace-d27","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Yuno","sameAs":"https://www.yuno.com/","logo":"https://logos.yubhub.co/yuno.com.png"},"x-apply-url":"https://jobs.lever.co/yuno/dc30ae7b-9c0f-426f-ae77-c58d9e4f6d6d","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Proven proactivity, technical acumen and the ability to lead initiatives and deliver projects.","Experience in defining and evolving data engineering standards, architectural guidelines and governance, ideally within a regulated environment.","Strong Python and SQL skills.","Hands-on experience with Spark or Flink in production.","DBT for data transformation."],"x-skills-preferred":["Airflow for orchestration.","Experience with Apache Hudi.","Experience with financial, transactional, or payment data."],"datePosted":"2026-04-17T13:11:45.302Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Europe"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Proven proactivity, technical acumen and the ability to lead initiatives and deliver projects., Experience in defining and evolving data engineering standards, architectural guidelines and governance, ideally within a regulated environment., Strong Python and SQL skills., Hands-on experience with Spark or Flink in production., DBT for data transformation., Airflow for orchestration., Experience with Apache Hudi., Experience with financial, transactional, or payment data."},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_886118d3-6a1"},"title":"Senior Data Engineer - Data Engineering","description":"<p>We believe that the way people interact with their finances will drastically improve in the next few years. We&#39;re dedicated to empowering this transformation by building the tools and experiences that thousands of developers use to create their own products.</p>\n<p>Plaid powers the tools millions of people rely on to live a healthier financial life. We work with thousands of companies like Venmo, SoFi, several of the Fortune 500, and many of the largest banks to make it easy for people to connect their financial accounts to the apps and services they want to use.</p>\n<p>The main goal of the DE team in 2024-25 is to build robust golden data sets to power our business goals of creating more insights-based products. Making data-driven decisions is key to Plaid&#39;s culture. To support that, we need to scale our data systems while maintaining correct and complete data.</p>\n<p>Data Engineers heavily leverage SQL and Python to build data workflows. We use tools like DBT, Airflow, Redshift, ElasticSearch, Atlanta, and Retool to orchestrate data pipelines and define workflows.</p>\n<p>We work with engineers, product managers, business intelligence, data analysts, and many other teams to build Plaid&#39;s data strategy and a data-first mindset.</p>\n<p>Our engineering culture is IC-driven -- we favor bottom-up ideation and empowerment of our incredibly talented team.</p>\n<p>We are looking for engineers who are motivated by creating impact for our consumers and customers, growing together as a team, shipping the MVP, and leaving things better than we found them.</p>\n<p>You will be in a high-impact role that will directly enable business leaders to make faster and more informed business judgments based on the datasets you build.</p>\n<p>You will have the opportunity to carve out the ownership and scope of internal datasets and visualizations across Plaid which is a currently unowned area that we intend to take over and build SLAs on.</p>\n<p>You will have the opportunity to learn best practices and up-level your technical skills from our strong DE team and from the broader Data Platform team.</p>\n<p>You will collaborate with and have strong and cross-functional partnerships with literally all teams at Plaid from Engineering to Product to Marketing/Finance etc.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Understanding different aspects of the Plaid product and strategy to inform golden dataset choices, design and data usage principles.</li>\n</ul>\n<ul>\n<li>Have data quality and performance top of mind while designing datasets</li>\n</ul>\n<ul>\n<li>Leading key data engineering projects that drive collaboration across the company.</li>\n</ul>\n<ul>\n<li>Advocating for adopting industry tools and practices at the right time</li>\n</ul>\n<ul>\n<li>Owning core SQL and python data pipelines that power our data lake and data warehouse.</li>\n</ul>\n<ul>\n<li>Well-documented data with defined dataset quality, uptime, and usefulness.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>4+ years of dedicated data engineering experience, solving complex data pipelines issues at scale.</li>\n</ul>\n<ul>\n<li>You&#39;ve have experience building data models and data pipelines on top of large datasets (in the order of 500TB to petabytes)</li>\n</ul>\n<ul>\n<li>You value SQL as a flexible and extensible tool, and are comfortable with modern SQL data orchestration tools like DBT, Mode, and Airflow.</li>\n</ul>\n<ul>\n<li>You have experience working with different performant warehouses and data lakes; Redshift, Snowflake, Databricks.</li>\n</ul>\n<ul>\n<li>You have experience building and maintaining batch and real-time pipelines using technologies like Spark, Kafka.</li>\n</ul>\n<ul>\n<li>You appreciate the importance of schema design, and can evolve an analytics schema on top of unstructured data.</li>\n</ul>\n<ul>\n<li>You are excited to try out new technologies. You like to produce proof-of-concepts that balance technical advancement and user experience and adoption.</li>\n</ul>\n<ul>\n<li>You like to get deep in the weeds to manage, deploy, and improve low-level data infrastructure.</li>\n</ul>\n<ul>\n<li>You are empathetic working with stakeholders. You listen to them, ask the right questions, and collaboratively come up with the best solutions for their needs while balancing infra and business needs.</li>\n</ul>\n<ul>\n<li>You are a champion for data privacy and integrity, and always act in the best interest of consumers.</li>\n</ul>\n<p><strong>Additional Information</strong></p>\n<p>Our mission at Plaid is to unlock financial freedom for everyone. To support that mission, we seek to build a diverse team of driven individuals who care deeply about making the financial ecosystem more equitable.</p>\n<p>We recognize that strong qualifications can come from both prior work experiences and lived experiences. We encourage you to apply to a role even if your experience doesn&#39;t fully match the job description.</p>\n<p>We are always looking for team members that will bring something unique to Plaid!</p>\n<p>Plaid is proud to be an equal opportunity employer and values diversity at our company. We do not discriminate based on race, color, national origin, ethnicity, religion or religious belief, sex (including pregnancy, childbirth, or related medical conditions), sexual orientation, gender, gender identity, gender expression, transgender status, sexual stereotypes, age, military or veteran status, disability, or other applicable legally protected characteristics.</p>\n<p>We also consider qualified applicants with criminal histories, consistent with applicable federal, state, and local laws.</p>\n<p>Plaid is committed to providing reasonable accommodations for candidates with disabilities in our recruiting process. If you need any assistance with your application or interviews due to a disability, please let us know at accommodations@plaid.com</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_886118d3-6a1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Plaid","sameAs":"https://plaid.com/","logo":"https://logos.yubhub.co/plaid.com.png"},"x-apply-url":"https://jobs.lever.co/plaid/022278b3-0943-44b3-a54b-1de421017589","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$190,800-$286,800 per year","x-skills-required":["SQL","Python","DBT","Airflow","Redshift","ElasticSearch","Atlanta","Retool","Spark","Kafka"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:52:06.845Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"SQL, Python, DBT, Airflow, Redshift, ElasticSearch, Atlanta, Retool, Spark, Kafka","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":190800,"maxValue":286800,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2324ce80-532"},"title":"Data Scientist - Network Value","description":"<p>We believe that the way people interact with their finances will drastically improve in the next few years. We&#39;re dedicated to empowering this transformation by building the tools and experiences that thousands of developers use to create their own products.</p>\n<p>Plaid powers the tools millions of people rely on to live a healthier financial life. We work with thousands of companies like Venmo, SoFi, several of the Fortune 500, and many of the largest banks to make it easy for people to connect their financial accounts to the apps and services they want to use.</p>\n<p>The Network Value Data Science team is helping Plaid build an industry leading fintech consumer network by increasing access to, authorization for, and usability of Plaid&#39;s User&#39;s financial footprints. We embed within product teams to support OKRs and help execute on product roadmaps. We translate ambiguous product questions into tractable analysis, serve as analytical thought partners throughout the org, identify opportunities to build better products, and champion a data-first decision making approach everywhere we go.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Perform ad-hoc and strategic analyses to uncover opportunities for improved business outcomes and translate complex questions into actionable analytics projects.</li>\n<li>Design and maintain scalable data models and dashboards that increase visibility into core systems and drive operational excellence.</li>\n<li>Build and iterate on machine learning prototypes to power insight-driven products and unlock new sources of customer and business value.</li>\n<li>Define and track OKRs that quantify progress toward key business goals, ensuring alignment and accountability across teams.</li>\n<li>Design and analyze experiments to guide product decisions and optimize feature launches.</li>\n<li>Champion a data-first culture by promoting analytical rigor and evidence-based decision-making across the organization.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>2+ years of experience as a Data Scientist or in a related analytics or data-focused role</li>\n<li>Strong track record of turning complex data into strategic insights and measurable business impact</li>\n<li>Proven ability to use experimentation, advanced analytics, and data storytelling to uncover opportunities that drive key product and business outcomes</li>\n<li>Strong technical foundation in SQL and Python for large-scale analysis, data modeling, and ML prototyping</li>\n<li>Experience developing and maintaining data pipelines and metrics frameworks using tools such as Airflow and dbt</li>\n<li>Background working with complex backend systems, ensuring data integrity, scalability, and operational reliability across platforms</li>\n<li>Skilled at partnering cross-functionally with product, engineering, and business teams to influence prioritization and strategy through clear, data-driven communication</li>\n</ul>\n<p><strong>Additional Information</strong></p>\n<p>Our mission at Plaid is to unlock financial freedom for everyone. To support that mission, we seek to build a diverse team of driven individuals who care deeply about making the financial ecosystem more equitable. We recognize that strong qualifications can come from both prior work experiences and lived experiences. We encourage you to apply to a role even if your experience doesn&#39;t fully match the job description.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2324ce80-532","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Plaid","sameAs":"https://plaid.com/","logo":"https://logos.yubhub.co/plaid.com.png"},"x-apply-url":"https://jobs.lever.co/plaid/18503c02-17a0-4c47-98c8-155b0b6ccc2a","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$176,400-$243,600 per year","x-skills-required":["SQL","Python","Machine Learning","Data Modeling","Data Pipelines","Metrics Frameworks","Airflow","dbt"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:52:02.474Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"SQL, Python, Machine Learning, Data Modeling, Data Pipelines, Metrics Frameworks, Airflow, dbt","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":176400,"maxValue":243600,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_586b9fef-509"},"title":"Senior Software Engineer - Network Enablement (Applied ML)","description":"<p>We believe that the way people interact with their finances will drastically improve in the next few years. We&#39;re dedicated to empowering this transformation by building the tools and experiences that thousands of developers use to create their own products.</p>\n<p>On this team, you will build and operate the ML infrastructure and product services that enable trust and intelligence across Plaid&#39;s network. You&#39;ll own feature engineering, offline training and batch scoring, online feature serving, and real-time inference so model outputs directly power partner-facing fraud &amp; trust products and bank intelligence features.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Embed model inference into Network Enablement product flows and decision logic (APIs, feature flags, backend flows).</li>\n<li>Define and instrument product + ML success metrics (fraud reduction, retention lift, false positives, downstream impact).</li>\n<li>Design and run experiments and rollout plans (backtesting, shadow scoring, A/B tests, feature-flagged releases) to validate product hypotheses.</li>\n<li>Build and operate offline training pipelines and production batch scoring for bank intelligence products.</li>\n<li>Ship and maintain online feature serving and low-latency model inference endpoints for real-time partner/bank scoring.</li>\n<li>Implement model CI/CD, model/version registry, and safe rollout/rollback strategies.</li>\n<li>Monitor model/data health: drift/regression detection, model-quality dashboards, alerts, and SLOs targeted to partner product needs.</li>\n<li>Ensure offline and online parity, data lineage, and automated validation / data contracts to reduce regressions.</li>\n<li>Optimize inference performance and cost for real-time scoring (batching, caching, runtime selection).</li>\n<li>Ensure fairness, explainability and PII-aware handling for partner-facing ML features; maintain auditability for compliance.</li>\n<li>Partner with platform and cross-functional teams to scale the ML/data foundation (graph features, sequence embeddings, unified pipelines).</li>\n<li>Mentor engineers and document team standards for ML productization and operations.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>Must-haves:</li>\n<li>Strong software engineering skills including systems design, APIs, and building reliable backend services (Go or Python preferred).</li>\n<li>Production experience with batch and streaming data pipelines and orchestration tools such as Airflow or Spark.</li>\n<li>Experience building or operating real-time scoring and online feature-serving systems, including feature stores and low-latency model inference.</li>\n<li>Experience integrating model outputs into product flows (APIs, feature flags) and measuring impact through experiments and product metrics.</li>\n<li>Experience with model lifecycle and operations: model registries, CI/CD for models, reproducible training, offline &amp; online parity, monitoring and incident response.</li>\n<li>Nice to have:</li>\n<li>Experience in fraud, risk, or marketing intelligence domains.</li>\n<li>Experience with feature-store products (Tecton / Chronon / Feast / internal) and unified pipelines.</li>\n<li>Experience with graph frameworks, graph feature engineering, or sequence embeddings.</li>\n<li>Experience optimizing inference at scale (Triton/ONNX/quantization, batching, caching).</li>\n</ul>\n<p><strong>Additional Information</strong></p>\n<p>Our mission at Plaid is to unlock financial freedom for everyone. To support that mission, we seek to build a diverse team of driven individuals who care deeply about making the financial ecosystem more equitable.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_586b9fef-509","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Plaid","sameAs":"https://plaid.com/","logo":"https://logos.yubhub.co/plaid.com.png"},"x-apply-url":"https://jobs.lever.co/plaid/43b1374d-5c5e-4b63-b710-a95e3cb76bbe","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$190,800-$286,800 per year","x-skills-required":["software engineering","systems design","APIs","backend services","Go","Python","batch and streaming data pipelines","orchestration tools","Airflow","Spark","real-time scoring","online feature-serving systems","feature stores","low-latency model inference","model outputs","product flows","experiments","product metrics","model lifecycle","operations","model registries","CI/CD","reproducible training","offline & online parity","monitoring","incident response"],"x-skills-preferred":["fraud","risk","marketing intelligence","feature-store products","unified pipelines","graph frameworks","graph feature engineering","sequence embeddings","inference at scale","Triton","ONNX","quantization","batching","caching"],"datePosted":"2026-04-17T12:51:26.228Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"software engineering, systems design, APIs, backend services, Go, Python, batch and streaming data pipelines, orchestration tools, Airflow, Spark, real-time scoring, online feature-serving systems, feature stores, low-latency model inference, model outputs, product flows, experiments, product metrics, model lifecycle, operations, model registries, CI/CD, reproducible training, offline & online parity, monitoring, incident response, fraud, risk, marketing intelligence, feature-store products, unified pipelines, graph frameworks, graph feature engineering, sequence embeddings, inference at scale, Triton, ONNX, quantization, batching, caching","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":190800,"maxValue":286800,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e503559e-cf7"},"title":"Senior Machine Learning Engineer","description":"<p><strong>Job Title: Senior Machine Learning Engineer</strong></p>\n<p><strong>Job Description:</strong></p>\n<p>Before 1965, it was extremely difficult and time-consuming to analyze complicated signals, like radio or images. You could solve it, but you had to throw a ton of compute at it. That all changed with the invention of the Fast Fourier transform, which could efficiently break that signal down into the frequencies that are a part of it.</p>\n<p>The Risk Onboarding team is working on efficiently reviewing customers’ applications without compromising on quality. We are the front line of defense for preventing money laundering and financial crimes, building systems to verify that someone is who they say they are and that we are allowed to do business with them.</p>\n<p><strong>About Us:</strong></p>\n<p>At Mercury, we craft an exceptional banking experience for startups. Our team is focused on ensuring our products create a safe environment that meets the needs of our customers, administrators, and regulators.</p>\n<p><strong>Job Responsibilities:</strong></p>\n<p>As part of this role, you will:</p>\n<ul>\n<li>Partner with data science &amp; engineering teams to design and deploy ML &amp; Gen AI microservices, primarily focusing on automating reviews</li>\n<li>Work with a full-stack engineering team to embed these services into the overall review experience, including human in the loop, escalations, and feeding human decisions back into the service</li>\n<li>Implement testing, observability, alerting, and disaster recovery for all services</li>\n<li>Implement tracing, performance, and regression testing</li>\n<li>Feel a strong sense of product ownership and actively seek responsibility – we often self-organize on small/medium projects, and we want someone who’s excited to help shape and build Mercury’s future</li>\n</ul>\n<p><strong>Ideal Candidate:</strong></p>\n<p>The ideal candidate for the role has:</p>\n<ul>\n<li>7+ years of experience in roles like machine learning engineering, data engineering, backend software engineering, and/or devops</li>\n<li>Expertise with:</li>\n</ul>\n<ul>\n<li>A full modern data stack: Snowflake, dbt, Fivetran, Airbyte, Dagster, Airflow</li>\n<li>SQL, dbt, Python</li>\n<li>OLAP / OLTP data modelling and architecture</li>\n<li>Key-value stores: Redis, dynamoDB, or equivalent</li>\n<li>Streaming / real-time data pipelines: Kinesis, Kafka, Redpanda</li>\n<li>API frameworks: FastAPI, Flask, etc.</li>\n<li>Production ML Service experience</li>\n<li>Working across full-stack development environment, with experience transferable to Haskell, React, and TypeScript</li>\n</ul>\n<p><strong>Total Rewards Package:</strong></p>\n<p>The total rewards package at Mercury includes base salary, equity (stock options/RSUs), and benefits. Our salary and equity ranges are highly competitive within the SaaS and fintech industry and are updated regularly using the most reliable compensation survey data for our industry. New hire offers are made based on a candidate’s experience, expertise, geographic location, and internal pay equity relative to peers.</p>\n<p><strong>Salary Range:</strong></p>\n<p>Our target new hire base salary ranges for this role are the following:</p>\n<ul>\n<li>US employees (any location): $200,700 - $250,900</li>\n<li>Canadian employees (any location): CAD 189,700 - 237,100</li>\n</ul>\n<p><strong>Diversity &amp; Belonging:</strong></p>\n<p>Mercury values diversity &amp; belonging and is proud to be an Equal Employment Opportunity employer. All individuals seeking employment at Mercury are considered without regard to race, color, religion, national origin, age, sex, marital status, ancestry, physical or mental disability, veteran status, gender identity, sexual orientation, or any other legally protected characteristic.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e503559e-cf7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Mercury","sameAs":"https://www.mercury.com/","logo":"https://logos.yubhub.co/mercury.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/mercury/jobs/5639559004","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$200,700 - $250,900 (US) | CAD 189,700 - 237,100 (Canada)","x-skills-required":["Snowflake","dbt","Fivetran","Airbyte","Dagster","Airflow","SQL","Python","OLAP / OLTP data modelling and architecture","Redis","dynamoDB","Kinesis","Kafka","Redpanda","FastAPI","Flask","Production ML Service experience","Haskell","React","TypeScript"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:45:16.566Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, New York, NY, Portland, OR, or Remote within Canada or United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Snowflake, dbt, Fivetran, Airbyte, Dagster, Airflow, SQL, Python, OLAP / OLTP data modelling and architecture, Redis, dynamoDB, Kinesis, Kafka, Redpanda, FastAPI, Flask, Production ML Service experience, Haskell, React, TypeScript","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":189700,"maxValue":250900,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8a8c0eb9-6e6"},"title":"Data Scientist, Product","description":"<p><strong>Job Title: Data Scientist, Product</strong></p>\n<p>This is the founding hire for product analytics at Hebbia. As a data scientist, you will define what our core product metrics are: what counts as an active user, what engagement actually means, what signals correlate with retention.</p>\n<p>This is not a dashboarding role. The goal is to shape product decisions with data, not just report on them. You will identify which workflows drive repeat usage, where users drop off, what features move engagement, and what differentiates power users from casual users across our enterprise customer base.</p>\n<p>The role sits at the intersection of analytics engineering, product analytics, and data science. You will build the infrastructure and do the analysis. Define the metrics, build the pipelines, create the dashboards, and use what you built to inform the roadmap.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Define and implement Hebbia&#39;s core product metrics from scratch: active users, engagement, retention, feature adoption, account health. Build the canonical definitions the entire company uses.</li>\n<li>Design and build the product analytics infrastructure: fact tables, clean data models, and the analytics layer that sits on top of our product data.</li>\n<li>Build and maintain executive and product dashboards that leadership and product teams use to make decisions.</li>\n<li>Write DAGs, transforms, and data pipelines that support analytics. Work with engineering to instrument the product so usage data is captured correctly.</li>\n<li>Analyze customer behavior across our B2B customer base: account-level usage patterns, workflow adoption, expansion signals, and churn risk indicators.</li>\n<li>Inform the product roadmap using data. Identify friction in user flows, surface feature adoption patterns, and highlight opportunities for product improvement.</li>\n<li>Partner with product managers and engineers to translate product questions into measurable data and structured experiments.</li>\n<li>Establish data quality standards and documentation so the metrics layer you build is trusted and maintained.</li>\n</ul>\n<p><strong>Who You Are</strong></p>\n<ul>\n<li>3+ years of experience in product analytics, analytics engineering, or data science at a B2B SaaS company or high-growth startup</li>\n<li>Strong in SQL and Python. You can write production-quality transforms, not just ad hoc queries.</li>\n<li>Experience with modern data stack tools: dbt, Airflow, Snowflake, BigQuery, or similar. You understand data modeling and warehouse architecture.</li>\n<li>You have built dashboards and reporting that product teams and leadership actually use to make decisions</li>\n<li>You understand B2B product analytics: account-level metrics, multi-user workflows, enterprise engagement patterns, and why B2B retention analysis is different from consumer</li>\n<li>You translate ambiguous product questions into structured analyses. You do not wait for someone to hand you a spec.</li>\n<li>Strong product intuition. You care about why users behave the way they do, not just what the numbers say.</li>\n<li>Clear communicator. You can present findings to engineers, product managers, and executives with equal effectiveness.</li>\n</ul>\n<p><strong>Compensation</strong></p>\n<p>The salary range for this position is set between $180,000 to $260,000. This range may be inclusive of several career levels at Hebbia and will be narrowed during the interview process based on the candidate’s experience and qualifications.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8a8c0eb9-6e6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Hebbia","sameAs":"https://hebbia.com/","logo":"https://logos.yubhub.co/hebbia.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/hebbia/jobs/4670090005","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$180,000 - $260,000","x-skills-required":["SQL","Python","dbt","Airflow","Snowflake","BigQuery","data modeling","warehouse architecture","product analytics","analytics engineering","data science"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:37:39.339Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City; San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"SQL, Python, dbt, Airflow, Snowflake, BigQuery, data modeling, warehouse architecture, product analytics, analytics engineering, data science","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":260000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2cf203a5-5c5"},"title":"Platform Engineer, Document Intelligence","description":"<p>About Hebbia</p>\n<hr>\n<p>The AI platform for investors and bankers that generates alpha and drives upside.</p>\n<p>Founded in 2020 by George Sivulka and backed by Peter Thiel and Andreessen Horowitz, Hebbia powers investment decisions for BlackRock, KKR, Carlyle, Centerview, and 40% of the world’s largest asset managers. Our flagship product, Matrix, delivers industry-leading accuracy, speed, and transparency in AI-driven analysis. It is trusted to help manage over $30 trillion in assets globally.</p>\n<p>We deliver the intelligence that gives finance professionals a definitive edge. Our AI uncovers signals no human could see, surfaces hidden opportunities, and accelerates decisions with unmatched speed and conviction. We do not just streamline workflows. We transform how capital is deployed, how risk is managed, and how value is created across markets.</p>\n<p>Hebbia is not a tool. Hebbia is the competitive advantage that drives performance, alpha, and market leadership.</p>\n<hr>\n<p>The Team</p>\n<hr>\n<p>The Document Intelligence team at Hebbia builds cutting-edge AI solutions that transform how users discover and interact with billions of private and public documents. Our products, including the Hebbia’s Browse application, enable intelligent document exploration, powerful search capabilities, and deep insights extraction. We focus on developing advanced data ingestion and search technologies that deliver intuitive, explainable, and highly responsive experiences. Working closely with customers, our team continuously iterates to address real-world challenges and drive impactful, data-driven decisions. Our goal is to empower users by seamlessly turning vast and complex document repositories into actionable intelligence.</p>\n<hr>\n<p>The Role</p>\n<hr>\n<p>Platform engineering at Hebbia is about excellent, scalable enablement. You are responsible for the core distributed systems that power billions of tokens across millions of dollars of AUM. You will be responsible for deploying efficient systems and building software tightly coupled with state-of-the-art infrastructure/system design. Hebbia’s edge is built on operating on the edge of the tokenomics curve and you will serve as a key contributor in this area. We value engineers who think on their feet, innovate and can solve for exponential scale.</p>\n<hr>\n<p>Responsibilities</p>\n<hr>\n<ul>\n<li>Own critical system components: Take complex requirements and turn them into robust, scaled solutions that solve real customer needs.</li>\n<li>Unlock O(1) universal indexing: Build and iterate on our high-scale document build system that enables constant time latency for indexing any content in the world, regardless of data volume.</li>\n<li>Drive performance optimization: Architect and implement performance-tuning solutions to ensure our systems operate efficiently at scale, minimizing latency and maximizing throughput across millions of documents.</li>\n<li>Mentor and guide: Provide technical leadership, mentorship, and guidance to junior engineers, fostering a culture of learning and growth.</li>\n</ul>\n<hr>\n<p>Who You Are</p>\n<hr>\n<ul>\n<li>Bachelor&#39;s or Master&#39;s degree in Computer Science, Data Science, Statistics, or a related field. A strong academic background with coursework in data structures, algorithms, and software development is preferred.</li>\n<li>5+ years software development experience at a venture-backed startup or top technology firm, with a focus on distributed systems and platform engineering.</li>\n<li>Proficiency in building backend and distributed systems using technologies such as Python, Java, or Go.</li>\n<li>Deep understanding of scalable system design, performance optimization, and resilience engineering.</li>\n<li>Extensive experience with cloud platforms (e.g., AWS).</li>\n<li>Working experience with one or more of the following: Kafka, ElasticSearch, PostgreSQL, and/or Redis.</li>\n<li>Knowledge of workflow orchestration and execution platforms like Airflow, Temporal or Prefect.</li>\n<li>Proven experience enabling observability patterns.</li>\n<li>Ability to analyze complex problems, propose innovative solutions, and effectively communicate technical concepts to both technical and non-technical stakeholders.</li>\n<li>Proven experience in leading software development projects and collaborating with cross-functional teams. Strong interpersonal and communication skills to foster a collaborative and inclusive work environment.</li>\n<li>Enthusiasm for continuous learning and professional growth. A passion for exploring new technologies, frameworks, and software development methodologies.</li>\n<li>Autonomous and excited about taking ownership over major initiatives.</li>\n</ul>\n<hr>\n<p>Bonuses:</p>\n<ul>\n<li>Experience building distributed systems leveraging technologies such as etcd or Apache Zookeeper.</li>\n<li>Frequent user of AI products, especially during the development lifecycle (i.e. Cursor, Claude Code, etc).</li>\n</ul>\n<hr>\n<p>Compensation</p>\n<hr>\n<p>The salary range for this role is $160,000 to $300,000. This range may be inclusive of several career levels at Hebbia and will be narrowed during the interview process based on the candidate’s experience and qualifications. Adjustments outside of this range may be considered for candidates whose qualifications significantly differ from those outlined in the job description.</p>\n<hr>\n<p>Life @ Hebbia</p>\n<hr>\n<ul>\n<li>PTO: Unlimited</li>\n<li>Insurance: Medical + Dental + Vision + 401K</li>\n<li>Eats: Catered lunch daily + doordash dinner credit if you ever need to stay late</li>\n<li>Parental leave policy: 3 months non-birthing parent, 4 months for birthing parent</li>\n<li>Fertility benefits: $15k lifetime benefit</li>\n<li>New hire equity grant: competitive equity package with unmatched upside potential</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2cf203a5-5c5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Hebbia","sameAs":"https://hebbia.com/","logo":"https://logos.yubhub.co/hebbia.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/hebbia/jobs/4584750005","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$160,000 to $300,000","x-skills-required":["backend and distributed systems","Python","Java","Go","scalable system design","performance optimization","resilience engineering","cloud platforms","AWS","Kafka","ElasticSearch","PostgreSQL","Redis","workflow orchestration and execution platforms","Airflow","Temporal","Prefect","observability patterns"],"x-skills-preferred":["etcd","Apache Zookeeper","AI products","Cursor","Claude Code"],"datePosted":"2026-04-17T12:37:13.599Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City; San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"backend and distributed systems, Python, Java, Go, scalable system design, performance optimization, resilience engineering, cloud platforms, AWS, Kafka, ElasticSearch, PostgreSQL, Redis, workflow orchestration and execution platforms, Airflow, Temporal, Prefect, observability patterns, etcd, Apache Zookeeper, AI products, Cursor, Claude Code","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":160000,"maxValue":300000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4475ebe1-e8a"},"title":"Data Engineering Intern","description":"<p>We&#39;re seeking a motivated and curious Data Engineering Intern to join our Data Platform team. This internship offers a unique opportunity to gain hands-on experience building and maintaining real data infrastructure within a fast-growing fintech environment.</p>\n<p>As a Data Engineering Intern, you&#39;ll collaborate on thoughtful projects and bring your fresh perspectives to impact our product and families. You&#39;ll assist in building and maintaining data pipelines using Airflow to orchestrate workflows that ingest, transform, and deliver data into Snowflake and Databricks.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Assist in building and maintaining data pipelines using Airflow to orchestrate workflows that ingest, transform, and deliver data into Snowflake and Databricks</li>\n<li>Support the design and implementation of data models in Snowflake that serve analytics, reporting, and ML use cases</li>\n<li>Help develop and maintain transformation logic using dbt, including writing models, tests, and documentation</li>\n<li>Contribute to data quality checks and validation processes to ensure accuracy, completeness, and timeliness of data</li>\n<li>Assist with infrastructure automation using Terraform to manage cloud resources in AWS</li>\n<li>Participate in troubleshooting data pipeline issues and investigating root causes alongside senior engineers</li>\n<li>Collaborate with data analysts, analytics engineers, and business stakeholders to understand requirements and contribute to technical solutions</li>\n<li>Help create and maintain documentation for data pipelines, data models, and infrastructure processes</li>\n<li>Participate in code reviews to develop best practices and learn from the team</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>A 3.0 GPA or higher</li>\n<li>Currently pursuing a Bachelor&#39;s or Master&#39;s degree in Computer Science, Data Science, Information Technology, or a related field</li>\n<li>Basic understanding of SQL and comfort manipulating data</li>\n<li>Interest in data engineering, data infrastructure, and/or analytics engineering</li>\n<li>Familiarity with Python for scripting, data processing, or automation (preferred)</li>\n<li>Basic understanding of cloud platforms, particularly AWS, is a plus</li>\n<li>Strong analytical thinking and problem-solving skills , especially comfort working through ambiguity</li>\n<li>Good communication and collaboration skills; able to work cross-functionally with technical and non-technical teammates</li>\n<li>Eagerness to take ownership of your work and ask thoughtful questions</li>\n</ul>\n<p>Learning Opportunities:</p>\n<ul>\n<li><p>You&#39;ll have the opportunity to gain experience with technologies including:</p>\n<ul>\n<li>Snowflake</li>\n<li>dbt (data build tool)</li>\n<li>Apache Airflow</li>\n<li>AWS (S3, Lambda, EC2, IAM)</li>\n<li>Databricks</li>\n<li>Terraform</li>\n<li>Fivetran</li>\n<li>Segment</li>\n</ul>\n</li>\n</ul>\n<p>This internship provides an excellent foundation for a career in data engineering, analytics engineering, or data architecture within the fintech industry.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4475ebe1-e8a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Greenlight","sameAs":"https://www.greenlight.com/","logo":"https://logos.yubhub.co/greenlight.com.png"},"x-apply-url":"https://jobs.lever.co/greenlight/b5d9d9b2-9d06-4db7-932c-30fd4a43825d","x-work-arrangement":"hybrid","x-experience-level":"intern","x-job-type":"internship","x-salary-range":null,"x-skills-required":["SQL","Python","Airflow","Snowflake","dbt","AWS","Databricks","Terraform","Fivetran","Segment"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:36:54.798Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Atlanta"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Finance","skills":"SQL, Python, Airflow, Snowflake, dbt, AWS, Databricks, Terraform, Fivetran, Segment"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_58df2f04-af4"},"title":"Data Engineer","description":"<p>We are looking for a Data Engineer to join our Data Platform team to partner with our product and business stakeholders across risk, operations, and other domains. As a Data Engineer, you will be responsible for building robust data pipelines and engineering foundations by ingesting data from disparate sources, ensuring data quality and consistency, and enabling better business decisions through reliable data infrastructure across core product areas.</p>\n<p>Your primary focus will be on building scalable data pipelines using Airflow to orchestrate data workflows that ingest, transform, and deliver data from various sources into Snowflake and Databricks. You will also design and implement data models in Snowflake that support analytics, reporting, and ML use cases with a focus on performance, reliability, and scalability.</p>\n<p>In addition, you will develop infrastructure as code using Terraform to automate and manage cloud resources in AWS, ensuring consistent and reproducible deployments. You will monitor data pipeline health and implement data quality checks to ensure accuracy, completeness, and timeliness of data as business needs evolve.</p>\n<p>You will also optimize data processing workflows to improve performance, reduce costs, and handle growing data volumes efficiently. Troubleshooting and resolving data pipeline issues, working through ambiguity to get to the root cause and implementing long-term fixes will be a key part of your role.</p>\n<p>As a Data Engineer, you will bridge gaps between data and the business by working with cross-functional teams across the US and India office to understand requirements and translate them into robust technical solutions. You will create comprehensive documentation on data pipelines, data models, and infrastructure, keeping documentation up to date and facilitating knowledge transfer across the team.</p>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>2+ years of data engineering experience with strong technical skills and the ability to architect scalable data solutions.</li>\n</ul>\n<ul>\n<li>Hands-on experience with Python for data processing, automation, and building data pipelines.</li>\n</ul>\n<ul>\n<li>Proficiency with workflow orchestration tools, preferably Airflow, including DAG development, task dependencies, and monitoring.</li>\n</ul>\n<ul>\n<li>Strong SQL skills and experience with cloud data warehouses like Snowflake, including performance optimization and data modeling.</li>\n</ul>\n<ul>\n<li>Experience with cloud platforms, preferably AWS (S3, Lambda, EC2, IAM, etc.), and understanding of cloud-based data architectures.</li>\n</ul>\n<ul>\n<li>Experience working cross-functionally with data analysts, analytics engineers, data scientists, and business stakeholders to understand requirements and deliver solutions.</li>\n</ul>\n<ul>\n<li>An ownership mentality – this engineer will be responsible for the reliability and performance of their data pipelines and expected to fully understand data flows, dependencies, and their implications on downstream users.</li>\n</ul>\n<p><strong>Nice to have:</strong></p>\n<ul>\n<li>Experience with dbt for transformation logic and analytics engineering workflows integrated with data pipelines.</li>\n</ul>\n<ul>\n<li>Familiarity with Databricks for large-scale data processing, including Spark optimization and Delta Lake.</li>\n</ul>\n<ul>\n<li>Experience with Infrastructure as Code (IaC) tools like Terraform for managing cloud resources and data infrastructure.</li>\n</ul>\n<ul>\n<li>Knowledge of data modeling concepts (e.g., dimensional modeling, star/snowflake schemas, slowly changing dimensions).</li>\n</ul>\n<ul>\n<li>Experience with CI/CD practices for data pipelines and automated testing frameworks.</li>\n</ul>\n<ul>\n<li>Experience with streaming data and real-time processing frameworks</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_58df2f04-af4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Greenlight","sameAs":"https://www.greenlight.com/","logo":"https://logos.yubhub.co/greenlight.com.png"},"x-apply-url":"https://jobs.lever.co/greenlight/e98d9733-8b8c-4ce4-997d-6cf14e35b2f3","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Python","SQL","Snowflake","Databricks","AWS","Terraform","data engineering","data pipelines","data modeling"],"x-skills-preferred":["dbt","Infrastructure as Code","CI/CD","streaming data","real-time processing"],"datePosted":"2026-04-17T12:36:30.660Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Airflow, Python, SQL, Snowflake, Databricks, AWS, Terraform, data engineering, data pipelines, data modeling, dbt, Infrastructure as Code, CI/CD, streaming data, real-time processing"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_eb26af8f-c1a"},"title":"Data Scientist","description":"<p>We are seeking a pragmatic, end-to-end Data Scientist who can operate across the full data lifecycle, from ingestion and modeling to productionizing key data systems. This is a high-impact, high-agency role which reports directly to the CTO. Modern AI-assisted development tools make this role possible, where the data scientist can now do real engineering, too.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Collaborate closely with other teams (Sales, Finance, Product, Marketing, and more) to translate problems and needs into action-oriented data solutions</li>\n<li>Design, build, and maintain data pipelines for reliable ingestion and transformation</li>\n<li>Rapidly prototype and iterate using AI coding tools to accelerate development and reduce toil</li>\n<li>Drive rigor and best practices, with a focus on data quality, consistency, and transparency</li>\n<li>Develop and deploy statistical models and machine learning, where appropriate</li>\n<li>Build clear, decision-oriented visualizations and dashboards for stakeholders across multiple departments</li>\n<li>Own selected production data systems: selection, orchestration, monitoring, and tuning</li>\n<li>Communicate and shepherd key data results and analysis to executives</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>Experience with B2B SaaS-relevant data, including Salesforce and financial metrics</li>\n<li>Strong communication skills and ability to work effectively across multiple departments and stakeholder groups</li>\n<li>Ownership mindset and ability to deliver end-to-end outcomes independently; must be a &quot;startup type&quot;</li>\n<li>Demonstrated ability to design data pipelines and work with imperfect, evolving data sources</li>\n<li>Sharp attention to data quality, including validation, anomaly detection, and root-cause analysis of inconsistencies</li>\n<li>Strong proficiency in Python and SQL; experience with modern data stack tools (e.g., dbt, Airflow, Spark, or equivalents, a plus)</li>\n<li>Experience with data visualization tools (e.g., Tableau, Looker, or similar)</li>\n<li>Some familiarity with infrastructure and related setup (databases, data warehouses, VMs)</li>\n<li>Knowledge of core machine learning concepts and when to apply them pragmatically</li>\n</ul>\n<p><strong>Initial Projects:</strong></p>\n<ul>\n<li>Build a likelihood-of-close model for Salesforce opportunities, which factors in relevant metadata and history</li>\n<li>Create a framework and initial implementation for an executive operational dashboard, working with a broad set of teams</li>\n<li>Define, validate, and implement key SaaS product-usage metrics</li>\n</ul>\n<p>As we grow, you will, too, with the broad scope of a software startup.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_eb26af8f-c1a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Forward Networks","sameAs":"https://www.forward.net/","logo":"https://logos.yubhub.co/forward.net.png"},"x-apply-url":"https://job-boards.greenhouse.io/forwardnetworks/jobs/7695301003","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$170,000 - $190,000","x-skills-required":["Python","SQL","data visualization","machine learning","data pipelines","data quality"],"x-skills-preferred":["dbt","Airflow","Spark","Tableau","Looker"],"datePosted":"2026-04-17T12:34:58.040Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Santa Clara, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, data visualization, machine learning, data pipelines, data quality, dbt, Airflow, Spark, Tableau, Looker","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":170000,"maxValue":190000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f0f321c2-15d"},"title":"Data Platform Engineer","description":"<p>At Anchorage Digital, we are building the world&#39;s most advanced digital asset platform for institutions to participate in crypto. Join the Data Platform team and build the Trusted Data Platform that powers Anchorage&#39;s transition to Data 3.0.</p>\n<p>You&#39;ll help shape the unified orchestration foundation, collaborate on governance-as-code patterns, and contribute to self-service frameworks that make quality and compliance automatic. We&#39;re moving from manual spreadsheets and theoretical architectures to automated control planes where every dataset is trusted, monitored, and traceable by default.</p>\n<p><strong>Technical Skills:</strong></p>\n<ul>\n<li>Collaborate on designing and implementing unified orchestration patterns (Dagster/Airflow) to replace legacy and fragmented scheduling</li>\n<li>Develop governance-as-code systems in partnership with the team that automatically apply policy tags, RLS, and access controls through an active control plane</li>\n</ul>\n<p><strong>Complexity and Impact of Work:</strong></p>\n<ul>\n<li>Help guide the technical design for platform capabilities like data contracts, automated quality gating, observability, and cost visibility</li>\n<li>Support the migration of workloads from legacy patterns to the modern platform, ensuring domain teams have clear paths and golden templates</li>\n</ul>\n<p><strong>Organizational Knowledge:</strong></p>\n<ul>\n<li>Partner with domain teams (Asset Data, Reporting &amp; Statements, Product teams) to understand their needs and design platform capabilities that enable their success</li>\n<li>Promote and support data mesh principles and dbt best practices, helping domain owners build and own their data products while platform ensures quality</li>\n</ul>\n<p><strong>Communication and Influence:</strong></p>\n<ul>\n<li>Promote data platform engineering best practices, developer experience, and &#39;Data as a Product&#39; principles across the engineering organization</li>\n<li>Contribute to architectural decisions and help establish engineering culture around reliability, cost efficiency, and operational excellence</li>\n</ul>\n<p><strong>You may be a fit for this role if you:</strong></p>\n<ul>\n<li>5-7+ years building data platforms or infrastructure: You bring experience helping design and operate modern data platforms that handle enterprise-scale workloads with quality, governance, and cost controls</li>\n<li>Strong dbt and SQL expertise: You&#39;re proficient with dbt and SQL, understand dbt Mesh, and have strong opinions on data modeling, testing, and documentation best practices</li>\n<li>Orchestration experience: You&#39;ve implemented production data orchestration with Airflow, Dagster, Prefect, or similar tools, and understand the trade-offs between different orchestration patterns</li>\n<li>Cloud data warehouse proficiency: You have strong experience with BigQuery, Snowflake, or Redshift, including query optimization, cost management, and security configurations</li>\n<li>Platform mindset: You think in terms of golden paths, reusable abstractions, and developer experience - you build systems that let others move fast safely</li>\n</ul>\n<p><strong>Although not a requirement, bonus points if:</strong></p>\n<ul>\n<li>Metadata and catalog experience: You&#39;ve worked with Atlan, Collibra, DataHub, or similar metadata platforms and understand active governance patterns</li>\n<li>Data observability tools: You&#39;ve implemented data quality monitoring with Great Expectations, Monte Carlo, Soda, or similar tools</li>\n<li>Infrastructure as code: You have experience with Terraform, Kubernetes, and modern DevOps practices for data infrastructure</li>\n<li>You&#39;re the kind of person who gets excited about declarative config, immutable infrastructure, and metrics dashboards showing cost-per-query trending down</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f0f321c2-15d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anchorage Digital","sameAs":"https://www.anchorage.co/","logo":"https://logos.yubhub.co/anchorage.co.png"},"x-apply-url":"https://jobs.lever.co/anchorage/8a325cd5-ef99-4f1e-bba8-7bb1fca64f12","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["dbt","SQL","Airflow","Dagster","Prefect","BigQuery","Snowflake","Redshift"],"x-skills-preferred":["Metadata and catalog experience","Data observability tools","Infrastructure as code"],"datePosted":"2026-04-17T12:24:40.602Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"dbt, SQL, Airflow, Dagster, Prefect, BigQuery, Snowflake, Redshift, Metadata and catalog experience, Data observability tools, Infrastructure as code"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_72eaaa6e-3c0"},"title":"Founding Engineer - Reporting & Statements","description":"<p>Join us as a founding engineer on our Reporting &amp; Statements team. You&#39;ll design the systems that power every financial report and statement we deliver from monthly reports to daily statements to custom client requests. We&#39;re building automated frameworks that guarantee accuracy and consistency for every number we send to clients.</p>\n<p><strong>Technical Skills:</strong></p>\n<ul>\n<li>Evolve our architecture from decentralized reporting scripts to a centralized, framework-based delivery system</li>\n<li>Build automated validation and reconciliation that lets us scale without adding manual oversight</li>\n</ul>\n<p><strong>Complexity and Impact of Work:</strong></p>\n<ul>\n<li>Design data models that become a trusted, shared source of truth for downstream product teams and external APIs</li>\n<li>Navigate complexity across multiple product data streams, applying consistent logic to all financial statements</li>\n</ul>\n<p><strong>Organizational Knowledge:</strong></p>\n<ul>\n<li>Work with Product and Foundations teams to standardize how we capture and represent financial data</li>\n<li>Create self-service frameworks so other teams can add new report types through configuration instead of code</li>\n</ul>\n<p><strong>Communication and Influence:</strong></p>\n<ul>\n<li>Listen to product stakeholders to stay ahead of scaling needs for client-facing data</li>\n<li>Help mature our engineering culture by advocating for and modeling &#39;Data as a Product&#39; principles and high-quality engineering standards</li>\n</ul>\n<p><strong>You may be a fit for this role if you:</strong></p>\n<ul>\n<li>7+ years building data systems: You have experience creating internal tools, frameworks, or engines that handle 10x scale</li>\n<li>Financial domain experience: You&#39;ve worked in fintech, banking, or other environments where numbers matter. You understand what a &#39;Statement of Record&#39; means and the precision it demands.</li>\n<li>Systems thinking: You consider the next 100 products, not just the current one. You value extensible systems over one-off pipelines.</li>\n<li>Solid technical foundation: You&#39;re proficient with Python (Pandas/Polars/Arrow) and SQL, with experience in BigQuery or similar cloud warehouses and modern orchestration tools like Airflow or Dagster.</li>\n</ul>\n<p><strong>Although not a requirement, bonus points if:</strong></p>\n<ul>\n<li>You&#39;ve been a data consumer: Prior experience as a financial or business analyst gives you the perspective to design truly usable data models.</li>\n<li>You care about performance: You enjoy making data move faster and cheaper, whether through ADBC, multiprocessing, vectorized operations, or other optimizations.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_72eaaa6e-3c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anchorage Digital","sameAs":"https://www.anchorage.co/","logo":"https://logos.yubhub.co/anchorage.co.png"},"x-apply-url":"https://jobs.lever.co/anchorage/5bcfc8f2-5f26-4f72-8ca7-f4b20ee7f7db","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","SQL","BigQuery","Airflow","Dagster","Pandas","Polars","Arrow"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:23:54.465Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, BigQuery, Airflow, Dagster, Pandas, Polars, Arrow"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3d849fbc-058"},"title":"Member of Product, Data Platform","description":"<p>At Anchorage Digital, we are building the world’s most advanced digital asset platform for institutions to participate in crypto.</p>\n<p>The Data Platform team is the backbone of Anchorage Digital&#39;s information infrastructure. As data becomes the lifeblood of every product, compliance workflow, and client-facing report we produce, this team is responsible for building and operating a unified, scalable, and reliable data platform that serves the entire organization.</p>\n<p>As a Data Platform Product Manager, you will own the strategy and execution for centralizing and formalizing the company&#39;s data infrastructure , spanning internal operational data, transaction and blockchain data, customer data, and external data sources.</p>\n<p>Your mission is to transform a fragmented data landscape into a single source of truth that powers mission-critical reporting, business insights, and downstream product experiences across every team at Anchorage.</p>\n<p>This is a force-multiplier role. Your work will elevate the quality, speed, and reliability of every product and team at the company.</p>\n<p>You will define the standards, build the platform, and create the foundation that enables Anchorage to scale with confidence.</p>\n<p>If you thrive at the intersection of complex data systems, cross-functional influence, and platform thinking, this is your opportunity to have outsized impact at a category-defining company in digital assets.</p>\n<p>Below, we define our Factors of Growth &amp; Impact to help Anchorage Villagers measure their impact and articulate feedback, coaching, and the rich learning that happens while exploring, developing, and mastering capabilities within and beyond the Member of Product, Data Platform role:</p>\n<p><strong>Technical Skills:</strong></p>\n<ul>\n<li>Own the detailed prioritization of the data platform roadmap, balancing foundational infrastructure work, new capabilities, and technical debt.</li>\n<li>Demonstrate deep strategic thinking in shaping the platform roadmap, considering the unique data challenges of digital assets, blockchain protocols, and regulated financial services.</li>\n<li>Deliver complex, cross-functional projects with multiple dependencies across engineering, analytics, compliance, and operations teams.</li>\n<li>Work closely with engineering and data science counterparts to drive product development processes, sprint planning, and architectural decisions.</li>\n<li>Ability to understand and reason about system architecture , including data warehousing, ETL/ELT pipelines, streaming vs. batch processing, and modern data stack components , and communicate clear requirements to engineering.</li>\n<li>Drive comprehensive go-to-market strategy for internal platform adoption, including defining success metrics, tracking KPIs around data quality and platform usage, and iterating based on data-driven insights.</li>\n</ul>\n<p><strong>Complexity and Impact of Work:</strong></p>\n<ul>\n<li>Lead and influence cross-functional teams while maintaining strong stakeholder relationships across the entire organization , from engineering to finance to compliance.</li>\n<li>Exercise independent decision-making and take full ownership of data platform strategy and execution.</li>\n<li>Contribute strategic insights that significantly impact company direction, operational efficiency, and product quality.</li>\n<li>Demonstrate platform leadership that elevates the performance and effectiveness of every team that depends on data.</li>\n</ul>\n<p><strong>Organizational Knowledge:</strong></p>\n<ul>\n<li>Develop deep understanding of Anchorage&#39;s business model, product suite, regulatory environment, and organizational structure.</li>\n<li>Build and maintain strong relationships with stakeholders across all departments to ensure the data platform serves the company&#39;s most critical needs.</li>\n<li>Navigate and improve organizational data practices to enhance efficiency, compliance, and decision-making.</li>\n<li>Drive company objectives through strategic data platform decisions and initiatives.</li>\n</ul>\n<p><strong>Communication and Influence:</strong></p>\n<ul>\n<li>Effectively influence and motivate teams across the organization to adopt platform standards and invest in data quality, even when those teams do not report to you.</li>\n<li>Enable cross-functional collaboration through clear, consistent communication about platform capabilities, timelines, and data governance expectations.</li>\n<li>Act as a thoughtful knowledge partner to senior leadership, translating complex data infrastructure topics into clear business impact.</li>\n<li>Proactively communicate platform goals, status updates, and data health metrics throughout the organization.</li>\n</ul>\n<p><strong>You may be a fit for this role if you:</strong></p>\n<ul>\n<li>5+ years of product management experience, with significant time spent on data platforms, data infrastructure, or data-intensive enterprise products.</li>\n<li>Proven experience building or scaling enterprise data platforms , including data warehousing, data lakes, ETL/ELT pipelines, or modern data stack tooling (e.g., Snowflake, Databricks, dbt, Airflow, Spark).</li>\n<li>Strong understanding of data modeling, data governance, and data quality frameworks.</li>\n<li>Experience working with diverse data types , including transactional data, customer data, financial data, and ideally blockchain or on-chain data.</li>\n<li>Track record of driving cross-functional alignment and adoption for internal platform products where you must influence without direct authority.</li>\n<li>Exceptional written and verbal communication skills, with the ability to convey complex data architecture concepts to both technical and non-technical audiences.</li>\n<li>Your empathy and adaptability not only complement others&#39; working styles but also embody our culture of curiosity, creativity, and shared understanding.</li>\n<li>You self describe as some combination of the following: creative, humble, ambitious, detail oriented, hard working, trustworthy, eager to learn, methodical, action oriented, and tenacious.</li>\n</ul>\n<p><strong>Although not a requirement, bonus points if you have:</strong></p>\n<ul>\n<li>You have hands-on experience with blockchain data indexing, onchain analytics, or crypto-native data infrastructure.</li>\n<li>You have built data platforms that serve both internal analytics consumers and external client-facing products (reports, statements, dashboards).</li>\n<li>You have experience supporting clients with data-related issues or concerns.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3d849fbc-058","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anchorage Digital","sameAs":"https://anchorage.com","logo":"https://logos.yubhub.co/anchorage.com.png"},"x-apply-url":"https://jobs.lever.co/anchorage/0e730f61-a2e4-4152-8277-3f6383cc69a6","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data platforms","data infrastructure","data-intensive enterprise products","data warehousing","data lakes","ETL/ELT pipelines","modern data stack tooling","Snowflake","Databricks","dbt","Airflow","Spark","data modeling","data governance","data quality frameworks","blockchain or on-chain data"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:18:21.529Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"United States"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data platforms, data infrastructure, data-intensive enterprise products, data warehousing, data lakes, ETL/ELT pipelines, modern data stack tooling, Snowflake, Databricks, dbt, Airflow, Spark, data modeling, data governance, data quality frameworks, blockchain or on-chain data"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3367a9d1-967"},"title":"Engineering Manager , Data Engineering Solutions","description":"<p>We&#39;re looking for a manager to drive the Data Engineering Solutions Team in solving high-impact, cutting-edge data problems. The ideal candidate will be someone that has built data pipelines for large scale volume, is deeply knowledgeable of Data Engineering tools including Airflow/Spark/Kafka/Flink, is empathetic, excels at building strong relationships, and collaborates effectively with other Stripe teams to understand their use cases and unlock new capabilities.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Deliver cutting-edge data pipelines that scale to users&#39; needs, focusing on reliability and efficiency.</li>\n<li>Lead and manage a team of ambitious, talented engineers, providing mentorship, guidance, and support to ensure their success.</li>\n<li>Drive the execution of key reporting initiatives for Stripe, overseeing the entire development lifecycle from planning to delivery while maintaining high standards of quality and timely completion.</li>\n<li>Collaborate with product managers and key leaders across the company to create a shared roadmap and drive adoption of canonical datasets and data warehouses, use golden paths, and ensure Stripes are using trustworthy data.</li>\n<li>Understand user needs and pain points to prioritize engineering work and deliver high-quality solutions that meet user needs.</li>\n<li>Provide hands-on technical leadership in architecture/design, vision/direction/requirements setting, and incident response processes for your reports.</li>\n<li>Foster a collaborative and inclusive work environment, promoting innovation, knowledge sharing, and continuous improvement within the team.</li>\n<li>Partner with our recruiting team to attract and hire top talent, and define the overall hiring strategies for your team.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3367a9d1-967","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com/","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7496118","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Spark","Kafka","Flink","Data Engineering","Team Management","Leadership","Communication","Problem-Solving"],"x-skills-preferred":["Iceberg","Change Data Capture","Hive Metastore","Pinot","Trino","AWS Cloud"],"datePosted":"2026-03-31T18:12:23.063Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Airflow, Spark, Kafka, Flink, Data Engineering, Team Management, Leadership, Communication, Problem-Solving, Iceberg, Change Data Capture, Hive Metastore, Pinot, Trino, AWS Cloud"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_91afbbda-8cd"},"title":"Backend Engineer, Data","description":"<p>We&#39;re looking for a talented Backend Engineer, Data to join our Data Foundations team. As a Backend Engineer, Data, you will design, develop, and own data pipelines, models, and products that power the Product, Data Science, and GTM functions. You will work with a variety of internal teams across Product, Data Science, and GTM to help them solve their data needs. Your work will provide visibility into how these stakeholders and the Data Foundations organization are performing and how we can deliver a better experience to Stripe&#39;s customers.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, develop, and own data pipelines, models, and products that power the Product, Data Science, and GTM functions</li>\n<li>Develop strong subject matter expertise and manage the SLAs for both data pipelines and full stack web applications that support these critical stakeholders</li>\n<li>Build and refine Stripe&#39;s data foundations - infrastructure, pipelines, and tools to enable various teams at Stripe - working with Scala, Spark, and Airflow</li>\n<li>Leverage LLM and Agents at scale to produce high-quality data on ambiguous problems</li>\n<li>Refine our existing data marts that help the GTM organization forecast the future potential performance of the business and reliably measure ongoing attainment toward targets</li>\n<li>Build data services that track key product metrics and measure the impact of different strategies employed by teams in the field</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li>Must have 6+ years of experience in a Software Engineering role, with a focus on building and maintaining data services, or data-intensive applications</li>\n<li>A strong engineering background and are interested in data</li>\n<li>Prior experience with writing and debugging data pipelines using a distributed data framework (Spark / Hadoop / Pig etc)</li>\n<li>An inquisitive nature in diving into data inconsistencies to pinpoint issues, and resolve deep rooted data quality issues</li>\n<li>Knowledge of a backend development language (such as Scala, Java, or Go) and strong SQL experience</li>\n<li>The ability to communicate cross-functionally, derive requirements and architect shared datasets</li>\n</ul>\n<p>Preferred Requirements:</p>\n<ul>\n<li>Experience creating and maintaining Data Marts to power business reporting needs</li>\n<li>Experience working with Product or GTM (Sales/Marketing) teams</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_91afbbda-8cd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com/","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/6865161","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Scala","Spark","Airflow","LLM","Agents","SQL","Java","Go"],"x-skills-preferred":["Data Marts","Product","GTM","Sales","Marketing"],"datePosted":"2026-03-31T18:01:28.144Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Canada"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Scala, Spark, Airflow, LLM, Agents, SQL, Java, Go, Data Marts, Product, GTM, Sales, Marketing"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1378ad18-3c9"},"title":"Data Team Leader","description":"<p>We are looking for an outstanding Data Team Leader to join our motivated engineering team. As a Data Team Leader, you will lead a dedicated group of data engineers, ensuring the successful implementation of innovative data pipelines and architectures.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Lead and coordinate a team of data engineers, ensuring delivery across multiple projects.</li>\n<li>Architect and implement scalable, high-performance data pipelines using Snowflake, dbt, and Airflow.</li>\n<li>Apply and guide others in using distributed systems and queueing technologies such as Celery, Redis, or equivalents.</li>\n<li>Own the end-to-end data lifecycle: ingestion, modeling, transformation, and delivery.</li>\n<li>Partner with cross-functional teams (product, analytics, DevOps) to meet business data needs.</li>\n<li>Enforce engineering guidelines, code quality, and performance standards.</li>\n<li>Conduct regular 1:1s, technical reviews, and provide mentorship to team members.</li>\n<li>Take initiative in capacity planning, hiring, and team scaling decisions.</li>\n</ul>\n<p>Requirements:</p>\n<ul>\n<li><p>5+ years of hands-on experience in data engineering.</p>\n</li>\n<li><p>2+ years of formal team leadership experience, including people management and project ownership.</p>\n</li>\n<li><p>Advanced knowledge of:</p>\n<ul>\n<li>Snowflake for warehousing and performance tuning.</li>\n<li>dbt for modular data modeling and testing.</li>\n<li>Apache Airflow (or similar workflow orchestrators).</li>\n<li>Distributed task and caching systems such as Celery, Redis, or similar technologies.</li>\n<li>Python, SQL, and shell scripting.</li>\n</ul>\n</li>\n<li><p>Experience with cloud platforms such as AWS, Azure, or GCP.</p>\n</li>\n<li><p>Strong grasp of software development guidelines, CI/CD, and data observability.</p>\n</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with real-time data streaming (e.g., Kafka).</li>\n<li>Familiarity with Terraform or other infrastructure-as-code tools.</li>\n<li>Prior experience in startup or high-growth environments.</li>\n<li>Exposure to BI platforms (e.g., Power BI, Looker, Tableau).</li>\n</ul>\n<p>Why Aristocrat?</p>\n<p>Aristocrat is a world leader in gaming content and technology, and a top-tier publisher of free-to-play mobile games. We deliver great performance for our B2B customers and bring joy to the lives of the millions of people who love to play our casino and mobile games.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1378ad18-3c9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Aristocrat","sameAs":"https://www.aristocrat.com/","logo":"https://logos.yubhub.co/aristocrat.com.png"},"x-apply-url":"https://aristocrat.wd3.myworkdayjobs.com/en-US/AristocratExternalCareersSite/job/Noida-UP-IN/Data-Team-Leader_R0020618","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Snowflake","dbt","Apache Airflow","Celery","Redis","Python","SQL","shell scripting","AWS","Azure","GCP","software development guidelines","CI/CD","data observability"],"x-skills-preferred":["real-time data streaming","Terraform","infrastructure-as-code tools","BI platforms"],"datePosted":"2026-03-10T12:13:54.011Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Noida, UP, IN"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Snowflake, dbt, Apache Airflow, Celery, Redis, Python, SQL, shell scripting, AWS, Azure, GCP, software development guidelines, CI/CD, data observability, real-time data streaming, Terraform, infrastructure-as-code tools, BI platforms"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_cb592721-c78"},"title":"Associate DevOps Engineer","description":"<p><strong>Associate DevOps Engineer991</strong></p>\n<p><strong>What we&#39;re all about.</strong></p>\n<p>Do you ever have the urge to do things better than the last time? We do. And it&#39;s this urge that drives us every day. Our environment of discovery and innovation means we&#39;re able to create deep and valuable relationships with our clients to create real change for them and their industries. It&#39;s what got us here – and it&#39;s what will make our future. At Quantexa, you&#39;ll experience autonomy and support in equal measures allowing you to form a career that matches your ambitions. 41% of our colleagues come from an ethnic or religious minority background. We speak over 20 languages across our 47 nationalities, creating a sense of belonging for all.</p>\n<p><strong>We&#39;re heading in one direction, the future. We&#39;d love you to join us.</strong></p>\n<p>At Quantexa we believe that people and organisations make better decisions when those decisions are put in context – we call this Contextual Decision Intelligence. Contextual Decision Intelligence is the new approach to data analysis that shows the relationships between people, places and organisations - all in one place - so you gain the context you need to make more accurate decisions, faster.</p>\n<p><strong>What will you be doing?</strong></p>\n<p>You&#39;ll be joining one of our DevOps teams in our R&amp;D department working on the Quantexa Cloud Platform and accompanying solutions. The platform is comprised of a landscape of low-maintenance, on-demand, and highly secure environments. Our environments host our software for our customers and partners to use, they also service a variety of internal use cases including underpinning the work of our R&amp;D teams to develop Quantexa Platform software.</p>\n<p>You&#39;ll be heavily involved with our cloud-based technical infrastructure, with responsibilities surrounding improving the availability and resilience of our platform, improving its usability and security, ensuring we stay at the forefront of technical innovation, and reducing toil across our estate.</p>\n<p>You will also work alongside our software engineering teams to leverage DevOps techniques to support our software release activities and work on unique cloud-based product offerings for our customers to use in their own DevOps processes on their own Cloud estate.</p>\n<p><strong>Our tech stack</strong></p>\n<ul>\n<li>A strong focus on Kubernetes &amp; GitOps, utilising tools like ArgoCD and Istio</li>\n<li>Infrastructure Management - CasC, IasC (Terraform, Docker, Ansible, Packer)</li>\n<li>Hybrid public Cloud, primarily GCP &amp; Azure, but also some AWS</li>\n<li>DevOps tooling/automation with the best tool for the job, commonly Bash, Python, Groovy, Golang</li>\n<li>Provisioning stack includes Elasticsearch, Spark, PostgreSQL, Valkey, Airflow, Kafka, etcd</li>\n<li>Log and metric aggregation with Fluentd, Prometheus, Grafana, Alertmanager</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<p><strong>We are looking for candidates who:</strong></p>\n<ul>\n<li>Take pride in designing, building and delivering high quality well engineered solutions to complex problems</li>\n<li>Take a big picture approach to solving problems, taking care to ensure that the solution works well within the wider system</li>\n<li>Commercial or non-commercial experience with programming/scripting/automation</li>\n<li>Good appreciation for information security principals</li>\n</ul>\n<p><strong>Experience in the following would be beneficial:</strong></p>\n<ul>\n<li>Experience with infrastructure management and general Linux administration</li>\n<li>Experience with software build and release engineering</li>\n<li>Exposure to a handful of the key parts of our tech stack listed above</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<p><strong>Why join Quantexa?</strong></p>\n<p>Our perks and quirks.</p>\n<p>What makes you Q will help you to realize your full potential, flourish and enjoy what you do, while being recognized and rewarded with our broad range of benefits.</p>\n<p>We offer:</p>\n<ul>\n<li>Competitive salary and Company Bonus</li>\n<li>Flexible working hours in a hybrid workplace &amp; free access to global WeWork locations &amp; events</li>\n<li>Pension Scheme with a company contribution of 6% (if you contribute 3%)</li>\n<li>25 days annual leave (with the option to buy up to 5 days) + birthday off!</li>\n<li>Work from Anywhere Scheme: Spend up to 2 months working outside of your country of employment over a rolling 12-month period</li>\n<li>Family: Enhanced Maternity, Paternity, Adoption, or Shared Parental Leave</li>\n<li>Private Healthcare with AXA</li>\n<li>EAP, Well-being Days, Gym Discounts</li>\n<li>Free Calm App Subscription #1 app for meditation, relaxation and sleep</li>\n<li>Workplace Nursery Scheme</li>\n<li>Team&#39;s Social Budget &amp; Company-wide Summer &amp; Winter Parties</li>\n<li>Tech &amp; Cycle-to-Work Schemes</li>\n<li>Volunteer Day off</li>\n<li>Dog-friendly Offices</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_cb592721-c78","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Quantexa","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/imLeMwxTKuwvDpxHC2mvRB/hybrid-associate-devops-engineer-in-london-at-quantexa","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Kubernetes","GitOps","ArgoCD","Istio","Infrastructure Management","CasC","IasC","Terraform","Docker","Ansible","Packer","Hybrid public Cloud","GCP","Azure","AWS","DevOps tooling/automation","Bash","Python","Groovy","Golang","Elasticsearch","Spark","PostgreSQL","Valkey","Airflow","Kafka","etcd","Fluentd","Prometheus","Grafana","Alertmanager"],"x-skills-preferred":[],"datePosted":"2026-03-09T17:03:44.848Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Kubernetes, GitOps, ArgoCD, Istio, Infrastructure Management, CasC, IasC, Terraform, Docker, Ansible, Packer, Hybrid public Cloud, GCP, Azure, AWS, DevOps tooling/automation, Bash, Python, Groovy, Golang, Elasticsearch, Spark, PostgreSQL, Valkey, Airflow, Kafka, etcd, Fluentd, Prometheus, Grafana, Alertmanager"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2b37c983-438"},"title":"Associate Data Engineer","description":"<p><strong><strong>Job Description</strong></strong></p>\n<p>You&#39;ll be joining the Applications team, which is an Engineering function within Quantexa&#39;s R&amp;D department that is focused on internally building real-world applications of the Quantexa Platform.</p>\n<p>This function enables demonstrations of the product and develops SaaS offerings, whilst also testing and refining new Platform features before they are deployed by clients. The function develops and releases its own tools that feed into these internal applications and are also packaged and released to help standardize and accelerate all Quantexa deployments.</p>\n<p>It encompasses four distinct sub-teams:</p>\n<p><strong><strong>Data Engineering Accelerators</strong></strong></p>\n<ul>\n<li>Developing Quantexa&#39;s libraries for cleansing, parsing and standardising data used in entity resolution</li>\n<li>Finding efficiency/performance improvements through big data testing and building performance tooling</li>\n<li>Owning best practices in entity resolution and network building</li>\n</ul>\n<p><strong><strong>Data Feeds</strong></strong></p>\n<ul>\n<li>Building standardised and reusable code for processing various third party/open source data sets</li>\n<li>Managing an internal data lake for the provision of this data by other teams for testing and analytics</li>\n<li>Owning general best practices for ingesting and processing data to get it ready for use in the Quantexa Platform, including pipelines and scheduling</li>\n</ul>\n<p><strong><strong>Demos</strong></strong></p>\n<ul>\n<li>Developing, deploying and maintaining all Quantexa demos, showcasing the different use cases for the Quantexa Platform</li>\n<li>Owning the Quantexa Trial platform, for prospective Quantexa clients to see the product in action using real data provided by Data Feeds</li>\n<li>Building tools to enable solution owners and sales to create their own custom demos</li>\n</ul>\n<p><strong><strong>SaaS</strong></strong></p>\n<ul>\n<li>Building Quantexa&#39;s emerging SaaS offering, a cloud hosted, standardized deployment of the Quantexa Platform</li>\n<li>Targeting mid-market banks in the US for Retail AML initially, providing them with a cost-effective Quantexa solution, then expanding in future to more use cases and geographies</li>\n<li>Implementing cutting edge features of the Quantexa Platform ensuring SaaS customers always on the latest and greatest of Quantexa</li>\n</ul>\n<p><strong><strong>Requirements</strong></strong></p>\n<ul>\n<li>Data processing/ETL pipelines</li>\n<li>Analysing and examining real and varied data</li>\n<li>Full stack development, but with a heavy focus on the data processing/ETL side</li>\n<li>Solving difficult problems with efficient, resilient, high impact code</li>\n<li>Working in the cloud with production-grade systems</li>\n<li>Defining best-practices and sharing expertise you’ve developed</li>\n<li>Working in a fast moving, Agile environment</li>\n<li>Growing and thriving within one of the UK’s fastest growing scale-ups</li>\n</ul>\n<p><strong><strong>Experience in the following would be beneficial:</strong></strong></p>\n<ul>\n<li>A strong coding background, ideally in Scala or otherwise in a relevant language that will allow you to learn Scala quickly (e.g. Java/Python)</li>\n<li>Big data, either from a software deployment/implementation or a data science perspective</li>\n<li>Working with big data technology, ideally Spark but others will also be useful such as Airflow or Elasticsearch</li>\n<li>Working in an Agile environment</li>\n<li>Building data processing pipelines for use in production batch systems, including either traditional ETL pipelines and/or analytics pipelines</li>\n<li>Manipulating data through cleansing, parsing, standardising etc, especially in relation to improving data quality/integrity</li>\n<li>Building and deploying SaaS products</li>\n</ul>\n<p><strong><strong>Benefits</strong></strong></p>\n<ul>\n<li>Competitive salary and Company Bonus</li>\n<li>Flexible working hours in a hybrid workplace &amp; free access to global WeWork locations &amp; events</li>\n<li>Pension Scheme with a company contribution of 6% (if you contribute 3%)</li>\n<li>25 days annual leave (with the option to buy up to 5 days) + birthday off!</li>\n<li>Work from Anywhere Scheme: Spend up to 2 months working outside of your country of employment over a rolling 12-month period</li>\n<li>Family: Enhanced Maternity, Paternity, Adoption, or Shared Parental Leave</li>\n<li>Private Healthcare with AXA</li>\n<li>EAP, Well-being Days, Gym Discounts</li>\n<li>Free Calm App Subscription #1 app for meditation, relaxation and sleep</li>\n<li>Workplace Nursery Scheme</li>\n<li>Team&#39;s Social Budget &amp; Company-wide Summer &amp; Winter Parties</li>\n<li>Tech &amp; Cycle-to-Work Schemes</li>\n<li>Volunteer Day off</li>\n<li>Dog-friendly Offices</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2b37c983-438","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Quantexa","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/ibgyYHfEjnWon7xfTj4nmA/associate-data-engineer-in-london-at-quantexa","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Data processing/ETL pipelines","Analysing and examining real and varied data","Full stack development","Solving difficult problems with efficient, resilient, high impact code","Working in the cloud with production-grade systems","Defining best-practices and sharing expertise you’ve developed","Working in a fast moving, Agile environment","Growing and thriving within one of the UK’s fastest growing scale-ups"],"x-skills-preferred":["A strong coding background, ideally in Scala or otherwise in a relevant language that will allow you to learn Scala quickly (e.g. Java/Python)","Big data, either from a software deployment/implementation or a data science perspective","Working with big data technology, ideally Spark but others will also be useful such as Airflow or Elasticsearch","Working in an Agile environment","Building data processing pipelines for use in production batch systems, including either traditional ETL pipelines and/or analytics pipelines","Manipulating data through cleansing, parsing, standardising etc, especially in relation to improving data quality/integrity","Building and deploying SaaS products"],"datePosted":"2026-03-09T17:03:01.548Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Data processing/ETL pipelines, Analysing and examining real and varied data, Full stack development, Solving difficult problems with efficient, resilient, high impact code, Working in the cloud with production-grade systems, Defining best-practices and sharing expertise you’ve developed, Working in a fast moving, Agile environment, Growing and thriving within one of the UK’s fastest growing scale-ups, A strong coding background, ideally in Scala or otherwise in a relevant language that will allow you to learn Scala quickly (e.g. Java/Python), Big data, either from a software deployment/implementation or a data science perspective, Working with big data technology, ideally Spark but others will also be useful such as Airflow or Elasticsearch, Working in an Agile environment, Building data processing pipelines for use in production batch systems, including either traditional ETL pipelines and/or analytics pipelines, Manipulating data through cleansing, parsing, standardising etc, especially in relation to improving data quality/integrity, Building and deploying SaaS products"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9c40b25b-28b"},"title":"FBS Senior Data Engineer (Airflow)","description":"<p>FBS – Farmer Business Services is part of Farmers operations with the purpose of building a global approach to identifying, recruiting, hiring, and retaining top talent. We believe that the foundation of every successful business lies in having the right people with the right skills. That is where we come in—helping Farmers build a winning team that delivers consistent and sustainable results.</p>\n<p>We don&#39;t have a local legal entity, so we&#39;ve partnered with Capgemini, which acts as the Employer of Record. Capgemini is responsible for managing local payroll and benefits.</p>\n<p>You can expect a solid and innovative company with a strong market presence, a dynamic, diverse, and multicultural work environment, leaders with deep market knowledge and strategic vision, and continuous learning and development.</p>\n<p>The new data platforms team will be our centralized shared services team supporting all data platforms such as Snowflake and Astronomer. They will be responsible for the strategy and implementation of these platforms as well as best practices for the business units to follow. In this case, the position is focused on Astronomer/Ariflow.</p>\n<p><strong>Key Responsibilities</strong></p>\n<ul>\n<li>Build and maintain automated data workflows and orchestrations using Apache Airflow</li>\n<li>Implement at least two major end-to-end data pipeline projects using Airflow</li>\n<li>Design and optimize complex DAGs for scalability, maintainability, and reliability</li>\n<li>Create reusable, parameterized, and modular Airflow components (operators, sensors, hooks) to streamline workflow development</li>\n<li>Ensure effective monitoring, alerting, and logging of Airflow DAGs for quick issue resolution</li>\n<li>Document workflows, solutions, and processes for team knowledge sharing and training</li>\n<li>Mentor and support other team members in Airflow usage and adoption</li>\n<li>Explain best practices, identify pros and cons, and communicate technical decisions to team members</li>\n<li>Develop reusable frameworks, leveraging reusable concepts for efficiency and scalability</li>\n<li>Implement and utilize reusable ecosystem components, including Python &amp; Apache Airflow, DynamoDB, Amazon RDS</li>\n<li>Develop reusable frameworks to enforce data governance and data quality standards</li>\n<li>CI/CD pipeline development using re-usable frameworks and Jenkins</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Between 4-6 years of experience in a similar role</li>\n<li>Bachelor&#39;s degree in IT, Information systems, Computer Science or a related field</li>\n<li>Insurance Experience (Desirable)</li>\n<li>Fluency in English</li>\n<li>Availability to work according to CST or PST time zones.</li>\n</ul>\n<p><strong>Technical Skills</strong></p>\n<ul>\n<li>Airflow (MUST) / Astronomer (PLUS) - Advanced (5 Years)</li>\n<li>Python - Advanced (4-6 Years) (MUST)</li>\n<li>Snowflake – Intermediate (MUST)</li>\n<li>DBT - Entry Level (PLUS)</li>\n<li>AWS Glue - Entry Level (PLUS)</li>\n<li>DynamoDB - Intermediate</li>\n<li>Amazon RDS - Intermediate</li>\n<li>Jenkins - Intermediate</li>\n</ul>\n<p><strong>Other Critical Skills</strong></p>\n<ul>\n<li>Work Independently</li>\n<li>Strategic Thinking</li>\n<li>Guide Others</li>\n<li>Documentation</li>\n<li>Explain best practices</li>\n<li>Communicate Technical Decisions</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<p>This position comes with a competitive compensation and benefits package.</p>\n<ul>\n<li>A competitive salary and performance-based bonuses.</li>\n<li>Comprehensive benefits package.</li>\n<li>Flexible work arrangements (remote and/or office-based).</li>\n<li>You will also enjoy a dynamic and inclusive work culture within a globally renowned group.</li>\n<li>Private Health Insurance.</li>\n<li>Paid Time Off.</li>\n<li>Training &amp; Development opportunities in partnership with renowned companies.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9c40b25b-28b","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/mzMboyMoUyGryzfFUD6uuZ/remote-fbs-senior-data-engineer-(airflow)-in-mexico-at-capgemini","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Python","Snowflake","DBT","AWS Glue","DynamoDB","Amazon RDS","Jenkins"],"x-skills-preferred":[],"datePosted":"2026-03-09T16:59:05.587Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Airflow, Python, Snowflake, DBT, AWS Glue, DynamoDB, Amazon RDS, Jenkins"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ba5e5f71-701"},"title":"FBS Associate Analytics Engineer","description":"<p>FBS Associate Analytics Engineer</p>\n<p>We are seeking an FBS Associate Analytics Engineer to join our team. As an FBS Associate Analytics Engineer, you will play a key role in transforming raw data into structured, high-quality datasets that are ready for analysis. You will work on low to moderately complex business problems, receiving coaching and guidance from data leadership. Your primary focus will be on end-to-end data workflow, including data ingestion, transformation, modeling, and validation to enable data-driven decision-making across the organization.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Emerging data infrastructure development with coaching and guidance: Pipeline Design and Development – Architects and builds scalable data pipelines using modern ETL (Extract, Load, Transform) tools and frameworks such as DBT (Data Build Tool), Apache Airflow, or similar.</li>\n<li>Automates data ingestion processes from various sources including databases, APIs, and third party services.</li>\n<li>Data Storage and Management - Designs and implements data warehousing solutions using platforms like Snowflake, Redshift, or BigQuery.</li>\n<li>Optimizes storage solutions for performance, cost efficiency, and scalability.</li>\n<li>Data Modeling - Develops and maintains logical and physical data models to support business analytics.</li>\n<li>Creates and manages dimensional models, star/snowflake schemas, and other data structures.</li>\n<li>Data Transformation - Transforms raw data into clean, organized, and analytics-ready datasets using SQL, Python, or other relevant languages.</li>\n<li>Data Quality Assurance - Conducts data validation and consistency checks to ensure the accuracy and reliability of data.</li>\n<li>Technology Stack - Utilizes modern data tools and technologies such as SQL, Python, dbt, Airflow, and cloud platforms like AWS, Azure, or GCP.</li>\n<li>Continuous Learning – Stays updated with the latest trends, best practices, and advancements in data engineering and analytics.</li>\n<li>Participates in professional development opportunities to enhance technical and analytical skills.</li>\n<li>Provides code as requirements for hardening and operationalization by technology with significant coaching, guidance, and feedback.</li>\n<li>Performs other duties as assigned.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>1+ year of experience working on a Data Environment</li>\n<li>Good Analytics mindset</li>\n<li>Knowledge in SQL</li>\n<li>Strong verbal communication and listening skills.</li>\n<li>Demonstrated written communication skills.</li>\n<li>Demonstrated analytical skills.</li>\n<li>Demonstrated problem solving skills.</li>\n<li>Effective interpersonal skills.</li>\n<li>Seeks to acquire knowledge in area of specialty.</li>\n<li>Possesses strong technical aptitude. Basic experience with SQL or similar, dimensional modeling, pipeline orchestration, building data pipelines to transform data, and BI visualizations.</li>\n<li>Python experience is a plus</li>\n</ul>\n<p>Benefits</p>\n<p>This position comes with a competitive compensation and benefits package.</p>\n<ul>\n<li>A competitive salary and performance-based bonuses.</li>\n<li>Comprehensive benefits package.</li>\n<li>Flexible work arrangements (remote and/or office-based).</li>\n<li>You will also enjoy a dynamic and inclusive work culture within a globally renowned group.</li>\n<li>Private Health Insurance.</li>\n<li>Paid Time Off.</li>\n<li>Training &amp; Development opportunities in partnership with renowned companies.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ba5e5f71-701","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/jaxxjRWH9XxkRbr1TCrPb5/remote-fbs-associate-analytics-engineer-in-mexico-at-capgemini","x-work-arrangement":"remote","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Python","DBT","Apache Airflow","Snowflake","Redshift","BigQuery","Data Modeling","Data Transformation","Data Quality Assurance","Cloud Platforms"],"x-skills-preferred":["Python experience"],"datePosted":"2026-03-09T16:55:09.881Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, DBT, Apache Airflow, Snowflake, Redshift, BigQuery, Data Modeling, Data Transformation, Data Quality Assurance, Cloud Platforms, Python experience"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2e1270db-bb7"},"title":"FBS Senior Data Engineer (Airflow)","description":"<p>FBS – Farmer Business Services is part of Farmers operations with the purpose of building a global approach to identifying, recruiting, hiring, and retaining top talent. We believe that the foundation of every successful business lies in having the right people with the right skills. That is where we come in—helping Farmers build a winning team that delivers consistent and sustainable results.</p>\n<p>We are looking for a Senior Data Engineer to join our new data platforms team, which will be our centralized shared services team supporting all data platforms such as Snowflake and Astronomer. The position is focused on Astronomer/Ariflow.</p>\n<p><strong>Key Responsibilities</strong></p>\n<ul>\n<li>Build and maintain automated data workflows and orchestrations using Apache Airflow</li>\n<li>Implement at least two major end-to-end data pipeline projects using Airflow</li>\n<li>Design and optimize complex DAGs for scalability, maintainability, and reliability</li>\n<li>Create reusable, parameterized, and modular Airflow components (operators, sensors, hooks) to streamline workflow development</li>\n<li>Ensure effective monitoring, alerting, and logging of Airflow DAGs for quick issue resolution</li>\n<li>Document workflows, solutions, and processes for team knowledge sharing and training</li>\n<li>Mentor and support other team members in Airflow usage and adoption</li>\n<li>Explain best practices, identify pros and cons, and communicate technical decisions to team members</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Between 4-6 years of experience in a similar role</li>\n<li>Bachelor&#39;s degree in IT, Information systems, Computer Science or a related field</li>\n<li>Insurance Experience (Desirable)</li>\n<li>Fluency in English</li>\n<li>Availability to work according to CST or PST time zones.</li>\n</ul>\n<p><strong>Technical Skills</strong></p>\n<ul>\n<li>Airflow (MUST) / Astronomer (PLUS) - Advanced (5 Years)</li>\n<li>Python - Advanced (4-6 Years) (MUST)</li>\n<li>Snowflake – Intermediate (MUST)</li>\n<li>DBT - Entry Level (PLUS)</li>\n<li>AWS Glue - Entry Level (PLUS)</li>\n<li>DynamoDB - Intermediate</li>\n<li>Amazon RDS - Intermediate</li>\n<li>Jenkins - Intermediate</li>\n</ul>\n<p><strong>Other Critical Skills</strong></p>\n<ul>\n<li>Work Independently</li>\n<li>Strategic Thinking</li>\n<li>Guide Others</li>\n<li>Documentation</li>\n<li>Explain best practices</li>\n<li>Communicate Technical Decisions</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<p>This position comes with a competitive compensation and benefits package.</p>\n<ul>\n<li>A competitive salary and performance-based bonuses.</li>\n<li>Comprehensive benefits package.</li>\n<li>Flexible work arrangements (remote and/or office-based).</li>\n<li>You will also enjoy a dynamic and inclusive work culture within a globally renowned group.</li>\n<li>Private Health Insurance.</li>\n<li>Paid Time Off.</li>\n<li>Training &amp; Development opportunities in partnership with renowned companies.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2e1270db-bb7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Capgemini","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/g6Kk9PeaSt9vgqEz7dTtY5/remote-fbs-senior-data-engineer-(airflow)-in-brazil-at-capgemini","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Python","Snowflake","DBT","AWS Glue","DynamoDB","Amazon RDS","Jenkins"],"x-skills-preferred":["Insurance Experience","Fluency in English"],"datePosted":"2026-03-09T16:54:51.948Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Airflow, Python, Snowflake, DBT, AWS Glue, DynamoDB, Amazon RDS, Jenkins, Insurance Experience, Fluency in English"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_591c2b45-829"},"title":"Data Ops, Vice President","description":"<p><strong>About this role</strong></p>\n<p>We are looking for an innovative hands-on technologist to run Data Operations for one of the largest global FinTech&#39;s. This is a new role that will transform how we manage and process high quality data at scale and reflects our commitment to invest in an Enterprise Data Platform to unlock our data strategy for BlackRock and our Aladdin Client Community.</p>\n<p><strong>Key responsibilities</strong></p>\n<p>The ideal candidate will be a high-energy, technology and data driven individual who has a track record of leading and doing the day to day operations.</p>\n<ul>\n<li>Ensure on time high quality data delivery with a single pane of glass for data pipeline observability and support</li>\n<li>Live and breathe best practices of data ops such as culture, processes and technology</li>\n<li>Partner cross-functionally to enhance existing data sets, eliminating manual inputs and ensuring high quality, and onboarding new data sets</li>\n<li>Lead change while ensuring daily operational excellence, quality, and control</li>\n<li>Build and maintain deep alignment with key internal partners on ops tooling and engineering</li>\n<li>Foster an agile collaborative culture which is creative open, supportive, and dynamic</li>\n</ul>\n<p><strong>Knowledge and Experience</strong></p>\n<ul>\n<li>8+ years’ experience in hands-on data operations including data pipeline monitoring and engineering</li>\n<li>Technical expert including experience with data processing, orchestration (Airflow) data ingestion, cloud-based databases/warehousing (Snowflake) and business intelligence tools</li>\n<li>Champion the adoption of AI-enabled capabilities across Data Operations, including anomaly detection, predictive issue management, and automated operational insights to improve resilience, scale, and client experience.</li>\n<li>The ability to operate and monitor large data sets through the data lifecycle, including the tooling and observability required to ensure data quality and control at scale</li>\n<li>Experience implementing, monitoring, and operating data pipelines that are fast, scalable, reliable, and accurate</li>\n<li>Understanding of modern-day data highways, the associated challenges, and effective controls</li>\n<li>Passionate about data platforms, data quality and everything data</li>\n<li>Practical and detailed oriented operations leader</li>\n<li>Inquisitive leader who will bring new ideas that challenge the status quo</li>\n<li>Ability to navigate a large, highly matrixed organization</li>\n<li>Strong presence with clients</li>\n</ul>\n<p><strong>Our benefits</strong></p>\n<p>To help you stay energized, engaged and inspired, we offer a wide range of employee benefits including: retirement investment and tools designed to help you in building a sound financial future; access to education reimbursement; comprehensive resources to support your physical health and emotional well-being; family support programs; and Flexible Time Off (FTO) so you can relax, recharge and be there for the people you care about.</p>\n<p><strong>Our hybrid work model</strong></p>\n<p>BlackRock’s hybrid work model is designed to enable a culture of collaboration and apprenticeship that enriches the experience of our employees, while supporting flexibility for all. Employees are currently required to work at least 4 days in the office per week, with the flexibility to work from home 1 day a week. Some business groups may require more time in the office due to their roles and responsibilities. We remain focused on increasing the impactful moments that arise when we work together in person – aligned with our commitment to performance and innovation. As a new joiner, you can count on this hybrid model to accelerate your learning and onboarding experience here at BlackRock.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_591c2b45-829","directApply":true,"hiringOrganization":{"@type":"Organization","name":"BlackRock","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/szVvonyUEuyEFVfTxyXWtS/data-ops%2C-vice-president-in-edinburgh-at-blackrock","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data operations","data pipeline monitoring","data ingestion","cloud-based databases","business intelligence tools","Airflow","Snowflake"],"x-skills-preferred":["AI-enabled capabilities","anomaly detection","predictive issue management","automated operational insights"],"datePosted":"2026-03-09T16:43:27.322Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Edinburgh, Scotland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"data operations, data pipeline monitoring, data ingestion, cloud-based databases, business intelligence tools, Airflow, Snowflake, AI-enabled capabilities, anomaly detection, predictive issue management, automated operational insights"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3487a0dd-b87"},"title":"Associate, Data Engineer","description":"<p><strong>Associate, Data Engineer at BlackRock</strong></p>\n<p>About this role</p>\n<p>BlackRock is looking for a data engineer to join the Digital Data Engineering team. In this role, you will help develop data integrations between BlackRock’s internal data systems and our external marketing technology platforms. You will work with business partners to develop data structures, build ETL pipelines, and implement appropriate data governance and monitoring.</p>\n<p>As part of BlackRock’s Digital organization, this role supports our mission to create AI-enabled, personalized and scalable marketing experiences. You will build the data foundations that power next generation digital platforms, audience personalization, and intelligent activation across a global ecosystem.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and build scalable data pipelines that support AI-enabled digital experiences, personalization, and marketing automation.</li>\n<li>Leverage AI-driven development and testing tools to increase engineering quality, speed, and reliability.</li>\n<li>Contribute to ongoing platform modernization efforts across Martech, content, analytics, and web ecosystems.</li>\n<li>Collaborate with cross-functional stakeholders to ensure data is structured and governed in ways that accelerate downstream personalization and analytics use cases.</li>\n<li>Architect and develop data solutions to bring new datasets into digital ecosystem including Private Markets data and product data.</li>\n</ul>\n<p><strong>Core Skills</strong></p>\n<ul>\n<li>You have flawless written and verbal communication and ability to gain buy-in on plans from a non-technical audience</li>\n<li>You have experience working with a broad set of stakeholders, including non-technical and non-quantitative people.</li>\n<li>You are comfortable using AI tools to enhance development workflows, such as prototyping, testing, documentation, and data validation.</li>\n<li>You have a strong desire to develop creatively and promote innovation.</li>\n<li>You&#39;re self-motivated and able to think big while also taking direction and feedback.</li>\n<li>You have excellent teamwork and collaboration skills.</li>\n</ul>\n<p><strong>Qualifications</strong></p>\n<ul>\n<li>3+ years’ experience in in SQL and Python, with experience in both RDBMS and Big Data structures. Existing experience with Snowflake-specific concepts is desirable.</li>\n<li>Familiarity with using AI-assisted development tools (e.g., code generation, code review, unit test development) to improve quality and delivery efficiency.</li>\n<li>ETL and pipeline development experience with Airflow and DBT is a plus.</li>\n<li>CI/CD experience with Azure and understanding of API frameworks is a plus.</li>\n<li>B.S. / M.S. degree in Computer Science, Engineering, or a related discipline.</li>\n<li>Knowledge of Marketing technology platforms is desirable, but not required (e.g., Eloqua/Marketo, web analytics platforms, customer data platforms).</li>\n<li>Relentless desire for understanding how processes work. Creativity in solving unconventional problems.</li>\n<li>Adaptability and resiliency when overcoming challenges.</li>\n</ul>\n<p><strong>Our benefits</strong></p>\n<p>To help you stay energized, engaged and inspired, we offer a wide range of employee benefits including: retirement investment and tools designed to help you in building a sound financial future; access to education reimbursement; comprehensive resources to support your physical health and emotional well-being; family support programs; and Flexible Time Off (FTO) so you can relax, recharge and be there for the people you care about.</p>\n<p><strong>Our hybrid work model</strong></p>\n<p>BlackRock’s hybrid work model is designed to enable a culture of collaboration and apprenticeship that enriches the experience of our employees, while supporting flexibility for all. Employees are currently required to work at least 4 days in the office per week, with the flexibility to work from home 1 day a week. Some business groups may require more time in the office due to their roles and responsibilities. We remain focused on increasing the impactful moments that arise when we work together in person – aligned with our commitment to performance and innovation. As a new joiner, you can count on this hybrid model to accelerate your learning and onboarding experience here at BlackRock.</p>\n<p><strong>About BlackRock</strong></p>\n<p>At BlackRock, we are all connected by one mission: to help more and more people experience financial well-being. Our clients, and the people they serve, are saving for retirement, paying for their children’s educations, buying homes and starting businesses. Their investments also help to strengthen the global economy: support businesses small and large; finance infrastructure projects that connect and power cities; and facilitate innovations that drive progress.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3487a0dd-b87","directApply":true,"hiringOrganization":{"@type":"Organization","name":"BlackRock","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/dgM5uMjA3xyRYgwF3u3x72/associate%2C-data-engineer-in-budapest-at-blackrock","x-work-arrangement":"hybrid","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Python","RDBMS","Big Data structures","Snowflake-specific concepts","AI-assisted development tools","code generation","code review","unit test development","ETL and pipeline development","Airflow","DBT","CI/CD experience","Azure","API frameworks","B.S. / M.S. degree in Computer Science","Engineering","or a related discipline"],"x-skills-preferred":["Marketing technology platforms","Eloqua/Marketo","web analytics platforms","customer data platforms"],"datePosted":"2026-03-09T16:41:54.882Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Budapest"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"SQL, Python, RDBMS, Big Data structures, Snowflake-specific concepts, AI-assisted development tools, code generation, code review, unit test development, ETL and pipeline development, Airflow, DBT, CI/CD experience, Azure, API frameworks, B.S. / M.S. degree in Computer Science, Engineering, or a related discipline, Marketing technology platforms, Eloqua/Marketo, web analytics platforms, customer data platforms"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_703876d0-bf6"},"title":"Senior Machine Learning Engineer: Ranking","description":"<p><strong>About Us</strong></p>\n<p>Constructor is a U.S. based company that develops a next-generation platform for search and discovery in ecommerce, built to optimize for metrics like revenue, conversion rate, and profit. Our search engine is entirely invented in-house utilizing transformers and generative LLMs, and we use its core and personalization capabilities to power everything from search itself to recommendations to shopping agents.</p>\n<p><strong>About the Team</strong></p>\n<p>The Ranking team, within the Machine Learning chapter, plays a central role in implementing algorithms that optimize our customers&#39; business KPIs like revenue and conversion rates. We focus on metrics over features, supplying our ranking algorithms with powerful capabilities that bring value to our customers.</p>\n<p><strong>Role Details</strong></p>\n<p><strong>Design and Develop ML-Based Ranking Solutions</strong></p>\n<p>As a Machine Learning Engineer on the Ranking team, your primary focus will be to enhance the quality of our ranking systems, ensuring that search, browse, and autocomplete experiences are highly relevant, personalized, and diverse. You will work on building state-of-the-art ranking algorithms that improve user experience and drive critical business metrics such as conversion, user engagement, and revenue growth.</p>\n<p><strong>Improve Ranking Quality</strong></p>\n<p>You will analyze ranking performance and identify gaps in search, browse, and autocomplete experiences, focusing on relevance, personalization, attractiveness, diversification, and other quality signals.</p>\n<p><strong>Innovate and Optimize Ranking Algorithms</strong></p>\n<p>You will proactively propose new machine learning models, algorithms, and features to advance the ranking pipeline, improve ranking quality, and meet evolving business needs.</p>\n<p><strong>Collaboration with Cross-Functional Teams</strong></p>\n<p>You will collaborate with technical and non-technical business partners to develop / update ranking functionalities (both within and outside the team)</p>\n<p><strong>Requirements</strong></p>\n<p><strong>Hard Skills</strong></p>\n<ul>\n<li>At least 4 years of experience with Python for machine learning and backend development</li>\n<li>At least 4 years of experience developing, deploying, and maintaining machine learning models with a strong focus on ranking systems for search, recommendations, or similar applications</li>\n<li>Experience in large-scale ML model training, evaluation, and optimization, with a focus on real-time inference and serving</li>\n<li>Experience with big data frameworks such as Spark for processing large datasets and integrating them into ML pipelines</li>\n<li>Proficiency in using tools like SQL, PySpark, Pandas, and other frameworks to extract, manipulate, and analyze data</li>\n<li>Experience with data pipeline orchestration tools like Airflow or Luigi to manage and automate workflows for ML training and signal delivery</li>\n<li>Experience working on ranking algorithms that optimize metrics such as relevance, conversion rates, personalization, user engagement, RPV is a plus</li>\n</ul>\n<p><strong>Soft Skills</strong></p>\n<ul>\n<li>Experience collaborating in cross-functional teams</li>\n<li>Experience leading projects to success</li>\n<li>Excellent English communication skills</li>\n<li>Enjoy helping others around you grow as developers and be successful</li>\n<li>Pick up new ideas and technologies quickly, love learning and talking to others about them</li>\n<li>Love to experiment and use data and customer feedback to drive decision making</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Unlimited vacation time</li>\n<li>Fully remote team</li>\n<li>Work from home stipend</li>\n<li>Apple laptops provided for new employees</li>\n<li>Training and development budget for every employee, refreshed each year</li>\n<li>Maternity &amp; Paternity leave for qualified employees</li>\n<li>Work with smart people who will help you grow and make a meaningful impact</li>\n<li>Base salary: $80k–$120k USD, depending on knowledge, skills, experience, and interview results</li>\n<li>Stock options</li>\n<li>Regular team offsites to connect and collaborate</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_703876d0-bf6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/C130DBB1DC","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$80k–$120k USD","x-skills-required":["Python","Machine learning","Backend development","Ranking systems","Search","Recommendations","Big data frameworks","Spark","SQL","PySpark","Pandas","Airflow","Luigi"],"x-skills-preferred":["Transformers","Generative LLMs","Personalization","User experience","Conversion","User engagement","Revenue growth"],"datePosted":"2026-03-09T10:59:16.198Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Machine learning, Backend development, Ranking systems, Search, Recommendations, Big data frameworks, Spark, SQL, PySpark, Pandas, Airflow, Luigi, Transformers, Generative LLMs, Personalization, User experience, Conversion, User engagement, Revenue growth","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":120000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_bb7bb8e9-e31"},"title":"Data Engineer - 12 Month TFT","description":"<p>We&#39;re looking for an experienced Data Engineer to join our team at Electronic Arts. As a Data Engineer, you will collaborate with the Marketing team to implement data strategies and develop complex ETL pipelines that support dashboards for promoting deeper understanding of our business.</p>\n<p>You will have experience developing and establishing scalable, efficient, automated processes for large-scale data analyses. You will also stay informed of the latest trends and research on all aspects of data engineering and analytics.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Design, implement and maintain efficient, scalable and robust data pipelines using cloud-native and open-source technologies</li>\n<li>Develop and optimize ETL/ELT processes to ingest, transform, and deliver data from diverse sources</li>\n<li>Automate deployment and monitoring of data workflows using CI/CD best practices</li>\n<li>Guide communications between our users and studio engineers to provide scalable end-to-end solutions</li>\n<li>Promote strategies to improve our data modelling, quality and architecture</li>\n<li>Participate in code reviews, mentor junior engineers, and contribute to team knowledge sharing</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li>4+ years relevant industry experience in a data engineering role and graduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field</li>\n<li>Proficiency in writing SQL queries and knowledge of cloud-based databases like Snowflake, Redshift, BigQuery or other big data solutions</li>\n<li>Experience in data modelling and tools such as dbt, ETL processes, and data warehousing</li>\n<li>Experience with at least one of the programming languages like Python, Java</li>\n<li>Experience with version control and code review tools such as Git</li>\n<li>Knowledge of latest data pipeline orchestration tools such as Airflow</li>\n<li>Experience with cloud platforms (AWS, GCP, or Azure) and infrastructure-as-code tools (e.g., Docker, Terraform, CloudFormation)</li>\n</ul>\n<p>Nice to Have:</p>\n<ul>\n<li>Experience in gaming and working with its telemetry data or data from similar sources</li>\n<li>Experience with big data platforms and technologies such as EMR, Databricks, Kafka, Spark, Iceberg</li>\n<li>Experience in developing engineering solutions based on near real-time/streaming dataset</li>\n<li>Exposure to AI/ML, MLOps concepts and collaboration with data science or AI teams.</li>\n</ul>\n<p>Pay Transparency - North America</p>\n<p>The ranges listed below are what EA in good faith expects to pay applicants for this role in these locations at the time of this posting. If you reside in a different location, a recruiter will advise on the applicable range and benefits. Pay offered will be determined based on a number of relevant business and candidate factors (e.g. education, qualifications, certifications, experience, skills, geographic location, or business needs).</p>\n<p>Pay Ranges: $100,000 - $139,500 CAD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_bb7bb8e9-e31","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Electronic Arts","sameAs":"https://jobs.ea.com","logo":"https://logos.yubhub.co/jobs.ea.com.png"},"x-apply-url":"https://jobs.ea.com/en_US/careers/JobDetail/Data-Engineer-12-month-TFT/212451","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"temporary","x-salary-range":"$100,000 - $139,500 CAD","x-skills-required":["SQL","cloud-based databases","data modelling","ETL processes","data warehousing","Python","Java","Git","Airflow","cloud platforms","infrastructure-as-code tools"],"x-skills-preferred":["gaming telemetry data","big data platforms","EMR","Databricks","Kafka","Spark","Iceberg","AI/ML","MLOps"],"datePosted":"2026-03-09T10:58:20.588Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver"}},"employmentType":"TEMPORARY","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, cloud-based databases, data modelling, ETL processes, data warehousing, Python, Java, Git, Airflow, cloud platforms, infrastructure-as-code tools, gaming telemetry data, big data platforms, EMR, Databricks, Kafka, Spark, Iceberg, AI/ML, MLOps","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":100000,"maxValue":139500,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0841fcf4-9ab"},"title":"Data Engineer SE - II","description":"<p>We are on a mission to rid the world of bad customer service by “mobilizing” the way help is delivered. Today’s consumers want an always-available customer service experience that leaves them feeling valued and respected.</p>\n<p>Helpshift helps B2B brands deliver this modern customer service experience through a mobile-first approach. We have changed how conversations take place, moving the conversation away from a slow, outdated email and desktop experience to an in-app chat experience that allows users to interact with brands in their own time.</p>\n<p>Through our market-leading AI-powered chatbots and automation, we help brands deliver instant and rapid resolutions. Because agents play a key role in delivering help, our platform gives agents superpowers with automation and AI that simply works.</p>\n<p><strong>About the Team</strong></p>\n<p>Consumers care first and foremost about having their time valued by brands. Brands need insights into their customer service operation to serve their consumers effectively. Such insights and analytics are delivered through various data products like in-app analytics dashboards and data-sharing integrations.</p>\n<p>The data platform team is responsible for designing, building, and maintaining the data infrastructure that enables such data and analytics products at scale. We build and manage data pipelines, databases, and other data structures to ensure that the data is reliable, accurate, and easily accessible.</p>\n<p>We also enable internal stakeholders with business intelligence and machine learning teams with data ops. This team manages the platform that handles 2 Million events per minute and processes 1+ terabytes of data daily.</p>\n<p><strong>About the Role</strong></p>\n<ul>\n<li>Building maintainable data pipelines both for data ingestion and operational analytics for data collected from 2 billion devices and 900M Monthly active users</li>\n<li>Building customer-facing analytics products that deliver actionable insights and data, easily detect anomalies</li>\n<li>Collaborating with data stakeholders to see what their data needs are and being a part of the analysis process</li>\n<li>Write design specifications, test, deployment, and scaling plans for the data pipelines</li>\n<li>Mentor people in the team &amp; organization</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>3+ years of experience in building and running data pipelines that scale for TBs of data</li>\n<li>Proficiency in high-level object-oriented programming language (Python or Java) is must</li>\n<li>Experience in Cloud data platforms like Snowflake and AWS, EMR/Athena is a must</li>\n<li>Experience in building modern data lakehouse architectures using Snowflake and columnar formats like Apache Iceberg/Hudi, Parquet, etc</li>\n<li>Proficiency in Data modeling, SQL query profiling, and data warehousing skills is a must</li>\n<li>Experience in distributed data processing engines like Apache Spark, Apache Flink, Datalfow/Apache Beam, etc</li>\n<li>Knowledge of workflow orchestrators like Airflow, Dasgter, etc is a plus</li>\n<li>Data visualization skills are a plus (PowerBI, Metabase, Tableau, Hex, Sigma, etc)</li>\n<li>Excellent verbal and written communication skills</li>\n<li>Bachelor’s Degree in Computer Science (or equivalent)</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Hybrid setup</li>\n<li>Worker&#39;s insurance</li>\n<li>Paid Time Offs</li>\n<li>Other employee benefits to be discussed by our Talent Acquisition team in India.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0841fcf4-9ab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Helpshift","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/D451DB2325","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","Snowflake","AWS","EMR/Athena","Apache Iceberg/Hudi","Parquet","Apache Spark","Apache Flink","Datalflow/Apache Beam","Airflow","Data modeling","SQL query profiling","data warehousing"],"x-skills-preferred":["PowerBI","Metabase","Tableau","Hex","Sigma"],"datePosted":"2026-03-09T10:52:21.768Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Pune, Maharashtra, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, Snowflake, AWS, EMR/Athena, Apache Iceberg/Hudi, Parquet, Apache Spark, Apache Flink, Datalflow/Apache Beam, Airflow, Data modeling, SQL query profiling, data warehousing, PowerBI, Metabase, Tableau, Hex, Sigma"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b1d4c773-5c5"},"title":"Analytics Engineer, Finance","description":"<p><strong>Compensation</strong></p>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>The Finance Data team is embedded within the CFO Org and is responsible for building internal data products that scale analytics across business teams and drive efficiencies in our daily operations. This team provides technical guidance on high-impact, scalable projects across Finance, and is the subject-matter expert in financial and transactional data that supports our Finance day-to-day operations.</p>\n<p><strong>About the Role</strong></p>\n<p>As an Analytics Engineer, you will be setting the foundation to scale analytics across our business functions and impart best data practices for a rapidly growing organization. We aspire to build the Finance team of the future.</p>\n<p>In addition, you will work collaboratively with key stakeholders in Finance and other business teams to understand their pain points and take the lead in proposing viable, future-proof solutions to resolve them. You will also autonomously lead your own projects that deliver business impact and help cultivate a mature data culture among Finance teams.</p>\n<p>We are looking for a seasoned engineer who has a proven track record of owning the entire data stack at high transaction volume companies, managing business critical ETL pipelines consumed by non-technical teams. As a generalist “fixer”, you may be deployed across several different Finance domains (e.g. Tax datamart, ERP migration, Procurement automation). For this role we need someone who excels in dynamic environments, adapts quickly to changing needs, and confidently navigates ambiguous or evolving requirements. If you&#39;re energized by solving technical problems without a playbook and comfortable wearing multiple hats, this role is for you! To clarify, you will <strong>not</strong> be responsible for training ML models and neither would we describe this role as ‘product analytics’.</p>\n<p>This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.</p>\n<p><strong>In this role, you will:</strong></p>\n<ul>\n<li>Understand the data needs of Finance teams, including Revenue, Tax, Procurement, Compute &amp; Infrastructure Accounting, Strategic Finance, and translate that scope into technical requirements</li>\n</ul>\n<ul>\n<li>Facilitate the development of data products and tools to for stakeholders to self-service and enable analytics to scale across the company</li>\n</ul>\n<ul>\n<li>Lead dimensional design - define, own, and maintain business facing data marts</li>\n</ul>\n<ul>\n<li>Be a cross-functional champion at upholding high data integrity standards and SLAs for the timely delivery of data</li>\n</ul>\n<ul>\n<li>Build and maintain insightful and reliable dashboards to track both operational and financial Metrics for the Executive team</li>\n</ul>\n<ul>\n<li>Contribute to the future roadmap of the Finance team from a data systems perspective</li>\n</ul>\n<ul>\n<li>Grow to be an expert in Finance Data and OpenAI’s data architecture</li>\n</ul>\n<p><strong>You might thrive in this role if you have:</strong></p>\n<ul>\n<li>7+ years of experience as an Analytics Engineer or in a similar role (Data Analyst or Data Engineer) with a proven track record in shipping canonical datasets</li>\n</ul>\n<ul>\n<li>Empathy towards non-developer stakeholders and their day-to-day pain points</li>\n</ul>\n<ul>\n<li>Strong proficiency in SQL for data transformation, comfort in at least one functional/OOP language such as Python or R</li>\n</ul>\n<ul>\n<li>Familiarity with managing distributed data stores (e.g. S3, Trino, Hive, Spark), and experience building multi-step ETL jobs coupled with orchestrating workflows (e.g. Airflow, Dagster)</li>\n</ul>\n<ul>\n<li>Experience in writing unit tests to validate data products and version control (e.g. GitHub, Stash)</li>\n</ul>\n<ul>\n<li>Expert at creating compelling data visualizations with dashboarding tools (e.g. Tableau, Looker or similar)</li>\n</ul>\n<ul>\n<li>Excellent communication skills and ability to present data-driven narratives in both verbal and written form to a non-technical audience</li>\n</ul>\n<ul>\n<li>Experience solving ambiguous problem statements in an early stage environment</li>\n</ul>\n<p><strong>You could be an especially great fit if you have:</strong></p>\n<ul>\n<li>Prior experience leading the development of an internal production tool, serving hundreds of cross-functional customers such as Billing Operations, Deal Desk or Go-to-Market teams</li>\n</ul>\n<ul>\n<li>Some frontend experience with React, TypeScript, Retool, Streamlit, or building web apps</li>\n</ul>\n<ul>\n<li>Good understanding of Spark and ability to write, debug, and optimize Spark jobs</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>\n<p>We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic.</p>\n<p>For additional information, please see [OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement](https://cdn.openai.com/policies/eeo-policy-statement.pdf).</p>\n<p>Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b1d4c773-5c5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/7cd50a19-65f2-4a52-89a2-512130e58c5c","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full time","x-salary-range":"$198K – $260K • Offers Equity","x-skills-required":["SQL","Python","R","S3","Trino","Hive","Spark","Airflow","Dagster","GitHub","Stash","Tableau","Looker"],"x-skills-preferred":["React","TypeScript","ReTool","Streamlit","Web development"],"datePosted":"2026-03-08T22:16:37.388Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, R, S3, Trino, Hive, Spark, Airflow, Dagster, GitHub, Stash, Tableau, Looker, React, TypeScript, ReTool, Streamlit, Web development","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":198000,"maxValue":260000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_18528dac-ae1"},"title":"Threat Collections Engineer","description":"<p><strong>About Anthropic</strong></p>\n<p>Anthropic&#39;s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems.</p>\n<p><strong>About the Role</strong></p>\n<p>We are looking for a Threat Collections Engineer to join our Threat Intelligence team. In this role, you will build the infrastructure that powers our threat discovery capabilities—integrating external data sources, developing detection systems for automated lead generation, and creating internal tooling that scales our investigators&#39; impact.</p>\n<p>This is a foundational engineering role on a small, high-impact team. You will take projects from proof-of-concept to production, work closely with investigators to understand their needs, and help scale what may become a multi-person collections function.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Build automated detection systems that use disparate signals to identify abusive behaviour.</li>\n<li>Take systems from idea to proof-of-concept to production-grade with appropriate monitoring, documentation, and maintenance processes</li>\n<li>Develop and maintain YARA rule infrastructure, including tools for writing, validating, and testing rules against real data</li>\n<li>Create integrations with external threat intelligence platforms (e.g. VirusTotal, Censys, Urlscan) via MCP servers to enable multi-source correlation during investigations</li>\n<li>Build data pipelines that ingest intelligence from RSS feeds, CTI news sources, and partner sharing, using Claude to extract TTPs and generate targeted hunting queries</li>\n<li>Develop behavioural analytics capabilities using DBT-based frameworks and create searchable audit logging infrastructure</li>\n<li>Establish feedback loops with investigators to tune detection systems and reduce false positives</li>\n<li>Scrape and normalise data from external sources to feed threat detection and enrichment workflows</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have strong coding proficiency in Python and SQL for building detection logic, data pipelines, and automation</li>\n<li>Have experience with data pipeline orchestration tools (Airflow, DBT, or similar)</li>\n<li>Have familiarity with threat intelligence concepts including IOCs, YARA rules, and threat correlation techniques</li>\n<li>Have experience integrating external APIs and building data ingestion systems</li>\n<li>Can translate investigator needs and workflows into technical requirements</li>\n<li>Are comfortable building v0 systems and iterating based on user feedback</li>\n<li>Have strong communication skills for working closely with non-engineering stakeholders</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Experience with threat intelligence sharing frameworks (e.g. MISP, STIX/TAXII)</li>\n<li>Background in cyber threat intelligence, security operations, or abuse detection</li>\n<li>Experience building MCP servers or similar tool integrations for AI systems</li>\n<li>Familiarity with web scraping and data extraction at scale</li>\n<li>Experience with behavioural analytics or anomaly detection systems</li>\n<li>Understanding of LLM capabilities and how to leverage them for automation</li>\n<li>A Top Secret Clearance</li>\n</ul>\n<p><strong>Deadline to apply:</strong></p>\n<p>None. Applications will be reviewed on a rolling basis.</p>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.** Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work. We think AI systems like the ones we&#39;re building have enormous social and ethical implications. We think this makes representation even more important, and we strive to include a range of diverse perspectives on our team.</strong></p>\n<p><strong>Your safety matters to us.** To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</strong></p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics as it does with computer science.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_18528dac-ae1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5074937008","x-work-arrangement":"remote","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$300,000 - $320,000 USD","x-skills-required":["Python","SQL","Airflow","DBT","YARA rules","Threat intelligence","API integration","Data ingestion","Web scraping","Data extraction"],"x-skills-preferred":["MISP","STIX/TAXII","Cyber threat intelligence","Security operations","Abuse detection","LLM capabilities","Automation"],"datePosted":"2026-03-08T13:53:41.541Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA, Washington, DC"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Airflow, DBT, YARA rules, Threat intelligence, API integration, Data ingestion, Web scraping, Data extraction, MISP, STIX/TAXII, Cyber threat intelligence, Security operations, Abuse detection, LLM capabilities, Automation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":300000,"maxValue":320000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1ace7478-7a2"},"title":"Staff+ Software Engineer, Data Infrastructure","description":"<p><strong>About the role</strong></p>\n<p>Data Infrastructure designs, operates, and scales secure, privacy-respecting systems that power data-driven decisions across Anthropic. Our mission is to provide data processing, storage, and access that are trusted, fast, and easy to use.</p>\n<p>We&#39;re looking for infrastructure engineers who thrive working at the intersection of data systems, security, and scalability. You&#39;ll tackle diverse challenges ranging from building financial reporting pipelines to architecting access control systems to ensuring cloud storage reliability. This role offers the opportunity to work directly with data scientists, analysts, and business stakeholders while diving deep into cloud infrastructure primitives.</p>\n<p><strong>Responsibilities:</strong></p>\n<p>Within Data Infra, you may be matched to critical business areas including:</p>\n<ul>\n<li><strong>Data Governance &amp; Access Control:</strong> Design and implement robust access control systems ensuring only authorized users can access sensitive data. Build infrastructure for permission management, audit logging, and compliance requirements. Work on IAM policies, ACLs, and security controls that scale across thousands of users and systems.</li>\n</ul>\n<ul>\n<li><strong>Financial Data Infrastructure:</strong> Build and maintain data pipelines and warehouses powering business-critical reporting. Ensure data integrity, accuracy, and availability for complex financial systems, including third party revenue ingestion pipelines; manage the external relationships as needed to drive upstream dependencies. Own the reliability of systems processing revenue, usage, and business metrics.</li>\n</ul>\n<ul>\n<li><strong>Cloud Storage &amp; Reliability:</strong> Architect disaster recovery, backup, and replication systems for petabyte-scale data. Ensure high availability and durability of data stored in cloud object storage (GCS, S3). Build systems that protect against data loss and enable rapid recovery.</li>\n</ul>\n<ul>\n<li><strong>Data Platform &amp; Tooling:</strong> Scale data processing infrastructure using technologies like BigQuery, BigTable, Airflow, dbt, and Spark. Optimize query performance, manage costs, and enable self-service analytics across the organization.</li>\n</ul>\n<p><strong>You might be a good fit if you:</strong></p>\n<ul>\n<li>Have 10+ years (not including internships or co-ops) of experience in a Software Engineer role, building data infrastructure, storage systems, or related distributed systems</li>\n</ul>\n<ul>\n<li>Have 3+ years (not including internships or co-ops) of experience leading large scale, complex projects or teams as an engineer or tech lead</li>\n</ul>\n<ul>\n<li>Can set technical direction for a team, not just execute within it</li>\n</ul>\n<ul>\n<li>Have deep experience with at least one of:</li>\n</ul>\n<ul>\n<li>Strong proficiency in programming languages like Python, Go, Java, or similar</li>\n</ul>\n<ul>\n<li>Experience with infrastructure-as-code (Terraform, Pulumi) and cloud platforms (GCP, AWS)</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Background in data warehousing, ETL/ELT pipelines, or analytics infrastructure</li>\n</ul>\n<ul>\n<li>Experience with Kubernetes, containerization, and cloud-native architectures</li>\n</ul>\n<ul>\n<li>Track record of improving data reliability, availability, or cost efficiency at scale</li>\n</ul>\n<ul>\n<li>Knowledge of column-oriented databases, OLAP systems, or big data processing frameworks</li>\n</ul>\n<ul>\n<li>Experience working in fintech, financial services, or highly regulated environments</li>\n</ul>\n<ul>\n<li>Security engineering background with focus on data protection and access controls</li>\n</ul>\n<p><strong>Technologies We Use:</strong></p>\n<ul>\n<li>Data: BigQuery, BigTable, Airflow, Cloud Composer, dbt, Spark, Segment, Fivetran</li>\n</ul>\n<ul>\n<li>Storage: GCS, S3</li>\n</ul>\n<ul>\n<li>Infrastructure: Terraform, Kubernetes, GCP, AWS</li>\n</ul>\n<ul>\n<li>Languages: Python, Go, SQL</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</p>\n<p><strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1ace7478-7a2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5114768008","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$405,000 - $485,000 USD","x-skills-required":["Python","Go","Java","Terraform","Pulumi","GCP","AWS","BigQuery","BigTable","Airflow","dbt","Spark","Segment","Fivetran","GCS","S3","Kubernetes","containerization","cloud-native architectures","data warehousing","ETL/ELT pipelines","analytics infrastructure","column-oriented databases","OLAP systems","big data processing frameworks","fintech","financial services","highly regulated environments","security engineering","data protection","access controls"],"x-skills-preferred":["data governance","access control","cloud storage","reliability","data platform","tooling","self-service analytics","data processing infrastructure","query performance","cost management"],"datePosted":"2026-03-08T13:52:03.469Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | Seattle, WA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, Java, Terraform, Pulumi, GCP, AWS, BigQuery, BigTable, Airflow, dbt, Spark, Segment, Fivetran, GCS, S3, Kubernetes, containerization, cloud-native architectures, data warehousing, ETL/ELT pipelines, analytics infrastructure, column-oriented databases, OLAP systems, big data processing frameworks, fintech, financial services, highly regulated environments, security engineering, data protection, access controls, data governance, access control, cloud storage, reliability, data platform, tooling, self-service analytics, data processing infrastructure, query performance, cost management","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":405000,"maxValue":485000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_6cc383e0-ff6"},"title":"ML Infrastructure Engineer, Safeguards","description":"<p><strong>About the role</strong></p>\n<p>We are seeking a Machine Learning Infrastructure Engineer to join our Safeguards organization, where you&#39;ll build and scale the critical infrastructure that powers our AI safety systems. You&#39;ll work at the intersection of machine learning, large-scale distributed systems, and AI safety, developing the platforms and tools that enable our safeguards to operate reliably at scale.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Design and build scalable ML infrastructure to support real-time and batch classifier and safety evaluations across our model ecosystem</li>\n<li>Build monitoring and observability tools to track model performance, data quality, and system health for safety-critical applications</li>\n<li>Collaborate with research teams to productionize safety research, translating experimental safety techniques into robust, scalable systems</li>\n<li>Optimize inference latency and throughput for real-time safety evaluations while maintaining high reliability standards</li>\n<li>Implement automated testing, deployment, and rollback systems for ML models in production safety applications</li>\n<li>Partner with Safeguards, Security, and Alignment teams to understand requirements and deliver infrastructure that meets safety and production needs</li>\n<li>Contribute to the development of internal tools and frameworks that accelerate safety research and deployment</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have 5+ years of experience building production ML infrastructure, ideally in safety-critical domains like fraud detection, content moderation, or risk assessment</li>\n<li>Are proficient in Python and have experience with ML frameworks like PyTorch, TensorFlow, or JAX</li>\n<li>Have hands-on experience with cloud platforms (AWS, GCP) and container orchestration (Kubernetes)</li>\n<li>Understand distributed systems principles and have built systems that handle high-throughput, low-latency workloads</li>\n<li>Have experience with data engineering tools and building robust data pipelines (e.g., Spark, Airflow, streaming systems)</li>\n<li>Are results-oriented, with a bias towards reliability and impact in safety-critical systems</li>\n<li>Enjoy collaborating with researchers and translating cutting-edge research into production systems</li>\n<li>Care deeply about AI safety and the societal impacts of your work</li>\n</ul>\n<p><strong>Strong candidates may have experience with:</strong></p>\n<ul>\n<li>Working with large language models and modern transformer architectures</li>\n<li>Implementing A/B testing frameworks and experimentation infrastructure for ML systems</li>\n<li>Developing monitoring and alerting systems for ML model performance and data drift</li>\n<li>Building automated labeling systems and human-in-the-loop workflows</li>\n<li>Experience in trust &amp; safety, fraud prevention, or content moderation domains</li>\n<li>Knowledge of privacy-preserving ML techniques and compliance requirements</li>\n<li>Contributing to open-source ML infrastructure projects</li>\n</ul>\n<p><strong>Deadline to apply:</strong></p>\n<p>None. Applications will be reviewed on a rolling basis.</p>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Education requirements: We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</li>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong></p>\n<p>Not all strong candidates will meet every single qualification as listed. Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work.</p>\n<p><strong>Your safety matters to us.</strong></p>\n<p>To protect yourself from potential scams, remember that Anthropic recruiters only contact you from @anthropic.com email addresses. In some cases, we may partner with vetted recruiting agencies who will identify themselves as working on behalf of Anthropic. Be cautious of emails from other domains. Legitimate Anthropic recruiters will never ask for money, fees, or banking information before your first day. If you&#39;re ever unsure about a communication, don&#39;t click any links—visit anthropic.com/careers directly for confirmed position openings.</p>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing the state of the art in AI safety and making a meaningful difference in the world.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_6cc383e0-ff6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4778843008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$320,000 - $405,000 USD","x-skills-required":["Python","PyTorch","TensorFlow","JAX","AWS","GCP","Kubernetes","Spark","Airflow","streaming systems"],"x-skills-preferred":["large language models","modern transformer architectures","A/B testing frameworks","experimentation infrastructure","monitoring and alerting systems","automated labeling systems","human-in-the-loop workflows","trust & safety","fraud prevention","content moderation domains","privacy-preserving ML techniques","compliance requirements"],"datePosted":"2026-03-08T13:46:05.401Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, PyTorch, TensorFlow, JAX, AWS, GCP, Kubernetes, Spark, Airflow, streaming systems, large language models, modern transformer architectures, A/B testing frameworks, experimentation infrastructure, monitoring and alerting systems, automated labeling systems, human-in-the-loop workflows, trust & safety, fraud prevention, content moderation domains, privacy-preserving ML techniques, compliance requirements","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":320000,"maxValue":405000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_63e3e256-1a7"},"title":"Senior Data Engineer","description":"<p><strong>Senior Data Engineer</strong></p>\n<p><strong>Location</strong></p>\n<p>London</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Location Type</strong></p>\n<p>Hybrid</p>\n<p><strong>Department</strong></p>\n<p>CommercialRevenue Operations</p>\n<p>Synthesia is the world&#39;s leading AI video platform for business, used by over 90% of the Fortune 100. Founded in 2017, the company is headquartered in London, with offices and teams across Europe and the US.</p>\n<p>As AI continues to shape the way we live and work, Synthesia develops products to enhance visual communication and enterprise skill development, helping people work better and stay at the centre of successful organisations.</p>\n<p>Following our recent Series E funding round, where we raised $200 million, our valuation stands at $4 billion. Our total funding exceeds $530 million from premier investors including Accel, NVentures (Nvidia&#39;s VC arm), Kleiner Perkins, GV, and Evantic Capital, alongside the founders and operators of Stripe, Datadog, Miro, and Webflow.</p>\n<p><strong>Senior Data Engineer</strong></p>\n<p>We&#39;re hiring a Senior Data Engineer to join Synthesia and take ownership of our core data systems. You&#39;ll be responsible for designing and maintaining scalable pipelines, optimising data models, and ensuring high data quality and governance standards.</p>\n<p><strong>What you&#39;ll do at Synthesia:</strong></p>\n<ul>\n<li>Architect and scale robust, end-to-end data pipelines that ingest and transform complex semi-structured and structured data into our Snowflake data warehouse.</li>\n</ul>\n<ul>\n<li>Own the evolution of our dbt project - implementing modular modelling patterns and other best practices to ensure a &#39;single source of truth&#39; for the entire organisation.</li>\n</ul>\n<ul>\n<li>Manage platform infrastructure in snowflake, AWS and other tools.</li>\n</ul>\n<ul>\n<li>Continuously optimise warehouse performance and cost by diagnosing bottlenecks, tuning inefficient queries, and improving how compute resources are used as we scale.</li>\n</ul>\n<ul>\n<li>Bridge the gap between experimental data science workflows and production, building the infrastructure and orchestration needed to deploy and monitor batch ML jobs.</li>\n</ul>\n<ul>\n<li>Drive best practices in data security, governance, and compliance, particularly with regards to AI.</li>\n</ul>\n<ul>\n<li>Partner with cross-functional stakeholders to understand data requirements and translate them into technical solutions.</li>\n</ul>\n<p><strong>What we&#39;re looking for:</strong></p>\n<ul>\n<li>5+ years of experience as a Data Engineer or in a closely related role, with a proven track record of building and operating production data systems.</li>\n</ul>\n<ul>\n<li>Experience working in an early-stage or scaling data function. You&#39;re comfortable taking ownership and wearing multiple hats when needed.</li>\n</ul>\n<ul>\n<li>Strong foundations in software engineering and data modelling best practices, with an ability to design systems that are maintainable, scalable, and easy for others to build on.</li>\n</ul>\n<ul>\n<li>Deep expertise in SQL, and solid experience using Python or similar languages to build data pipelines, tooling, and orchestration (Airflow).</li>\n</ul>\n<ul>\n<li>Hands on experience managing cloud infrastructure using infrastructure-as-code (e.g. Terraform) on AWS, GCP, or similar platforms.</li>\n</ul>\n<ul>\n<li>A pragmatic approach to data platform design, with an eye for performance, cost efficiency, and operational reliability.</li>\n</ul>\n<ul>\n<li>Excellent communication skills: you can work effectively with technical and non-technical stakeholders to gather requirements, explain trade-offs and communicate data team needs.</li>\n</ul>\n<ul>\n<li>A product-oriented mindset, with an understanding of how data can shape decision making and accelerate company growth.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_63e3e256-1a7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Synthesia","sameAs":"https://www.synthesia.io/","logo":"https://logos.yubhub.co/synthesia.io.png"},"x-apply-url":"https://jobs.ashbyhq.com/synthesia/46650970-494a-4d4b-ab4b-75c2a3b06daf","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Python","Airflow","Terraform","AWS","GCP","Snowflake","dbt"],"x-skills-preferred":[],"datePosted":"2026-03-06T18:32:20.724Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Airflow, Terraform, AWS, GCP, Snowflake, dbt"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4b563c21-dd0"},"title":"Software Engineer, Data Infrastructure","description":"<p><strong>Software Engineer, Data Infrastructure</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Department</strong></p>\n<p>Applied AI</p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$185K – $385K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>Data Platform at OpenAI owns the foundational data stack powering critical product, research, and analytics workflows. We operate some of the largest Spark compute fleets in production; design, and build data lakes and metadata systems on Iceberg and Delta with a vision toward exabyte-scale architecture; run high throughput streaming platforms on Kafka and Flink; provide orchestration with Airflow; and support ML feature engineering tooling such as Chronon. Our mission is to deliver reliable, secure, and efficient data access at scale and accelerate intelligent, AI assisted data workflows.</p>\n<p><strong>About the Role</strong></p>\n<p>This role focuses on building and operating data infrastructure that supports massive compute fleets and storage systems, designed for high performance and scalability. You’ll help design, build, and operate the next generation of data infrastructure at OpenAI. You will scale and harden big data compute and storage platforms, build and support high-throughput streaming systems, build and operate low latency data ingestions, enable secure and governed data access for ML and analytics, and design for reliability and performance at extreme scale.</p>\n<p>You will take full lifecycle ownership: architecture, implementation, production operations, and on-call participation.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design, build, and maintain data infrastructure systems such as distributed compute, data orchestration, distributed storage, streaming infrastructure, machine learning infrastructure while ensuring scalability, reliability, and security</li>\n</ul>\n<ul>\n<li>Ensure our data platform can scale by orders of magnitude while remaining reliable and efficient</li>\n</ul>\n<ul>\n<li>Accelerate company productivity by empowering your fellow engineers &amp; teammates with excellent data tooling and systems</li>\n</ul>\n<ul>\n<li>Collaborate with product, research and analytics teams to build the technical foundations capabilities that unlock new features and experiences</li>\n</ul>\n<ul>\n<li>Own the reliability of the systems you build, including participation in an on-call rotation for critical incidents</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>4+ years in data infrastructure engineering OR</li>\n</ul>\n<ul>\n<li>4+ years in infrastructure engineering with a strong interest in data</li>\n</ul>\n<ul>\n<li>Take pride in building and operating scalable, reliable, secure systems</li>\n</ul>\n<ul>\n<li>Are comfortable with ambiguity and rapid change</li>\n</ul>\n<ul>\n<li>Have an intrinsic desire to learn and fill in missing skills, and an equally strong talent for sharing learnings clearly and concisely with others</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of human diversity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4b563c21-dd0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/f763c6b3-5167-4a67-b691-4c3fa2c44156","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"$185K – $385K • Offers Equity","x-skills-required":["data infrastructure engineering","infrastructure engineering","Spark","Kafka","Flink","Airflow","Chronon","Iceberg","Delta","Terraform","distributed systems"],"x-skills-preferred":["machine learning","data science","cloud computing","containerization","DevOps"],"datePosted":"2026-03-06T18:23:49.267Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data infrastructure engineering, infrastructure engineering, Spark, Kafka, Flink, Airflow, Chronon, Iceberg, Delta, Terraform, distributed systems, machine learning, data science, cloud computing, containerization, DevOps","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":185000,"maxValue":385000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c873a489-0dc"},"title":"Data Engineer, Analytics","description":"<p><strong>Data Engineer, Analytics</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Department</strong></p>\n<p>Applied AI</p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$230K – $385K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the team</strong></p>\n<p>The Applied team works across research, engineering, product, and design to bring OpenAI’s technology to consumers and businesses.</p>\n<p>We seek to learn from deployment and distribute the benefits of AI, while ensuring that this powerful tool is used responsibly and safely. Safety is more important to us than unfettered growth.</p>\n<p><strong>About the role</strong></p>\n<p>We&#39;re seeking a Data Engineer to take the lead in building our data pipelines and core tables for OpenAI. These pipelines are crucial for powering analyses, safety systems that guide business decisions, product growth, and prevent bad actors. If you&#39;re passionate about working with data and are eager to create solutions with significant impact, we&#39;d love to hear from you. This role also provides the opportunity to collaborate closely with the researchers behind ChatGPT and help them train new models to deliver to users. As we continue our rapid growth, we value data-driven insights, and your contributions will play a pivotal role in our trajectory. Join us in shaping the future of OpenAI!</p>\n<p><strong>In this role, you will:</strong></p>\n<ul>\n<li>Design, build and manage our data pipelines, ensuring all user event data is seamlessly integrated into our data warehouse.</li>\n</ul>\n<ul>\n<li>Develop canonical datasets to track key product metrics including user growth, engagement, and revenue.</li>\n</ul>\n<ul>\n<li>Work collaboratively with various teams, including, Infrastructure, Data Science, Product, Marketing, Finance, and Research to understand their data needs and provide solutions.</li>\n</ul>\n<ul>\n<li>Implement robust and fault-tolerant systems for data ingestion and processing.</li>\n</ul>\n<ul>\n<li>Participate in data architecture and engineering decisions, bringing your strong experience and knowledge to bear.</li>\n</ul>\n<ul>\n<li>Ensure the security, integrity, and compliance of data according to industry and company standards.</li>\n</ul>\n<p><strong>You might thrive in this role if you:</strong></p>\n<ul>\n<li>Have 3+ years of experience as a data engineer and 8+ years of any software engineering experience(including data engineering).</li>\n</ul>\n<ul>\n<li>Proficiency in at least one programming language commonly used within Data Engineering, such as Python, Scala, or Java.</li>\n</ul>\n<ul>\n<li>Experience with distributed processing technologies and frameworks, such as Hadoop, Flink and distributed storage systems (e.g., HDFS, S3).</li>\n</ul>\n<ul>\n<li>Expertise with any of ETL schedulers such as Airflow, Dagster, Prefect or similar frameworks.</li>\n</ul>\n<ul>\n<li>Solid understanding of Spark and ability to write, debug and optimize Spark code.</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c873a489-0dc","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/fc5bbc77-a30c-4e7a-9acc-8a2e748545b4","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$230K – $385K • Offers Equity","x-skills-required":["Python","Scala","Java","Hadoop","Flink","HDFS","S3","Airflow","Dagster","Prefect","Spark"],"x-skills-preferred":[],"datePosted":"2026-03-06T18:20:01.101Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Scala, Java, Hadoop, Flink, HDFS, S3, Airflow, Dagster, Prefect, Spark","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":230000,"maxValue":385000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_783aed89-627"},"title":"Data Scientist - PhD Intern (Short Term)","description":"<p><strong>[2026] Data Scientist - PhD Intern (Short Term)</strong></p>\n<p>San Mateo, CA, United States Data Science &amp; Analytics ID: 5750</p>\n<p>Every day, tens of millions of people come to Roblox to explore, create, play, learn, and connect with friends in 3D immersive digital experiences– all created by our global community of developers and creators.</p>\n<p>At Roblox, we’re building the tools and platform that empower our community to bring any experience that they can imagine to life. Our vision is to reimagine the way people come together, from anywhere in the world, and on any device.</p>\n<p>A career at Roblox means you’ll be working to shape the future of human interaction, solving unique technical challenges at scale, and helping to create safer, more civil shared experiences for everyone.</p>\n<p>The Data Science &amp; Analytics organization&#39;s mission is to increase our speed, frequency and acumen of making decisions at scale by instilling a data-influenced approach to building products. We cover a wide area of the data spectrum including analytical data engineering, product analytics, causal inference, economics, statistical modeling and machine learning. Aligned and partnering with product verticals, we use this extensive toolbelt to discover new opportunities and unmet use cases, influence and shape the product roadmap and prioritization, build data products and measure impact on our community of players and developers.</p>\n<p>This is a temporary, part-time position requiring no more than 20 hours per week _for a 3-month duration, with possibility to extend._</p>\n<p><strong>Teams Hiring for this role:</strong></p>\n<ul>\n<li><strong>Foundation AI:</strong> Our AI evaluation team focuses on generating high-quality models and consistently improving our evaluation models.</li>\n<li><strong>Safety:</strong> Managing account relationships and the real-time morphing of linguistic mapping.</li>\n<li><strong>Economy:</strong> Drive creator success and growth by exploring marketplace structure and pricing.</li>\n</ul>\n<p><strong>You Will:</strong></p>\n<ul>\n<li>Collaborate with data scientists and engineers to research and develop advanced data analytics, causal inference, experiment design and machine learning solutions to power the business and product innovations.</li>\n<li>Conduct in-depth research to address complex data-related challenges.</li>\n<li>Work on projects that have a real impact on our products, services, and business strategy.</li>\n<li>Apply your work to expedite product innovations, including in-experience experiments, friend recommendations, and dynamic resource allocation for experience servers</li>\n<li>Present your findings and recommendations to both technical and non-technical stakeholders.</li>\n</ul>\n<p><strong>You Have:</strong></p>\n<ul>\n<li>Possessing or pursuing a PhD degree in a quantitative field such as Statistics, Applied Math, Computer Science, Economics, or Computational Social Science, Operations Research, Computer Engineering, Electrical Engineering.</li>\n<li>At least 1 year of experience doing causal inference or machine learning or experiment design via research or prior internship.</li>\n<li>Proficiency in one or more programming languages (e.g., SQL, Python or R)</li>\n<li>Proficiency in big data query/processing languages and tools such as SQL, Hive, Spark, or Airflow.</li>\n<li>Passion for applying scientific rigor to advance dynamic consumer products.</li>\n<li>Experience in developing production solutions is a plus.</li>\n<li>Experience with ML modeling</li>\n</ul>\n<p>You may redact age, date of birth, and dates of attendance/graduation from your resume if you prefer.</p>\n<p>For roles that are based at our headquarters in San Mateo, CA: The starting base pay for this position is as shown below. The actual base pay is dependent upon a variety of job-related factors such as professional background, training, work experience, location, business needs and market demand. Therefore, in some circumstances, the actual hourly rate could fall outside of this expected range. This pay range is subject to change and may be modified in the future. _Please note that not all benefits shown on this page are applicable to internship opportunities._</p>\n<p>Hourly Pay Range</p>\n<p>$64—$64 USD</p>\n<p>Roles that are based in an office are onsite Tuesday, Wednesday, and Thursday, with optional presence on Monday and Friday (unless otherwise noted).</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_783aed89-627","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Roblox","sameAs":"https://careers.roblox.com","logo":"https://logos.yubhub.co/careers.roblox.com.png"},"x-apply-url":"https://careers.roblox.com/jobs/7540083","x-work-arrangement":"onsite","x-experience-level":"entry","x-job-type":"internship","x-salary-range":"$64—$64 USD","x-skills-required":["SQL","Python","R","Hive","Spark","Airflow","Statistics","Applied Math","Computer Science","Economics","Computational Social Science","Operations Research","Computer Engineering","Electrical Engineering"],"x-skills-preferred":["Machine Learning","Causal Inference","Experiment Design"],"datePosted":"2026-03-06T14:20:53.011Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Mateo, CA"}},"employmentType":"INTERN","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, R, Hive, Spark, Airflow, Statistics, Applied Math, Computer Science, Economics, Computational Social Science, Operations Research, Computer Engineering, Electrical Engineering, Machine Learning, Causal Inference, Experiment Design"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f7c94e9c-5ab"},"title":"Member of Technical Staff, Software Engineer","description":"<p><strong>Summary</strong></p>\n<p>Microsoft AI are looking for a talented Member of Technical Staff, Software Engineer to join their MAI SuperIntelligence team in Zürich, Switzerland. This role sits at the heart of strategic decision-making, turning market data into actionable insights for a company that&#39;s revolutionising AI technology. You&#39;ll work directly with leadership to shape the company&#39;s direction in the AI market.</p>\n<p><strong>About the Role</strong></p>\n<p>As a Member of Technical Staff, Software Engineer, you will design and build core platform services for scalable training and evaluation, including cluster orchestration, job scheduling, data and compute pipelines, and artifact management. You will standardize containerized workflows by maintaining Docker images, CI/CD, and runtime configurations; advocate for best practices in security, reproducibility, and cost efficiency. You will implement end-to-end observability and operations through metrics, tracing, logging, dashboard development, monitoring, and automated alerts for model training and platform health (using Prometheus, Grafana, OpenTelemetry). You will architect and operate services on Azure cloud platforms, managing infrastructure-as-code (Terraform/Helm), secrets, networking, and storage. You will enhance developer experience by creating tools, CLIs, and portals that simplify job submission, metrics analysis, and experiment management for generalist software engineering and research teams.</p>\n<p><strong>Accountabilities</strong></p>\n<ul>\n<li>Design and build core platform services for scalable training and evaluation, including cluster orchestration, job scheduling, data and compute pipelines, and artifact management.</li>\n<li>Standardize containerized workflows by maintaining Docker images, CI/CD, and runtime configurations; advocate for best practices in security, reproducibility, and cost efficiency.</li>\n</ul>\n<p><strong>The Candidate we&#39;re looking for</strong></p>\n<p><strong>Experience:</strong></p>\n<ul>\n<li>Strong software engineering background building reliable, scalable production systems (Python preferred).</li>\n</ul>\n<p><strong>Technical skills:</strong></p>\n<ul>\n<li>Hands-on experience supporting large-scale ML / LLM training, evaluation, or experimentation infrastructure.</li>\n<li>Operating GPU-heavy workloads in cloud environments using Docker and Kubernetes (scheduling, utilization, isolation).</li>\n<li>Designing and running data / compute pipelines and orchestration (e.g., Airflow, Argo) with object storage (Azure Blob / S3).</li>\n</ul>\n<p><strong>Personal attributes:</strong></p>\n<ul>\n<li>Building secure, reproducible platforms using CI/CD, infrastructure-as-code (Terraform, Helm), container security, and secrets management.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Competitive salary and benefits package.</li>\n<li>Opportunity to work with a talented team of engineers and researchers.</li>\n<li>Access to cutting-edge technology and resources.</li>\n<li>Flexible work arrangements, including remote work options.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f7c94e9c-5ab","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Microsoft AI","sameAs":"https://microsoft.ai","logo":"https://logos.yubhub.co/microsoft.ai.png"},"x-apply-url":"https://microsoft.ai/job/member-of-technical-staff-software-engineer-mai-superintelligence-team/","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"Competitive salary and benefits package","x-skills-required":["Strong software engineering background","Python","Docker","Kubernetes","Airflow","Argo","Azure Blob","S3"],"x-skills-preferred":["CI/CD","Terraform","Helm","Container security","Secrets management"],"datePosted":"2026-03-06T07:32:22.031Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Zürich, Switzerland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Strong software engineering background, Python, Docker, Kubernetes, Airflow, Argo, Azure Blob, S3, CI/CD, Terraform, Helm, Container security, Secrets management"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2902359a-64d"},"title":"Member of Technical Staff, Infrastructure Data & Analytics","description":"<p><strong>Summary</strong></p>\n<p>Microsoft AI are looking for a talented Member of Technical Staff, Infrastructure Data &amp; Analytics to join their MAI SuperIntelligence Team. This role sits at the heart of strategic decision-making, turning raw telemetry into trusted, decision-quality insights on utilization, capacity, readiness, and efficiency. You&#39;ll work directly with leadership to shape the company&#39;s direction in the Superintelligence space.</p>\n<p><strong>About the Role</strong></p>\n<p>As a Member of Technical Staff, Infrastructure Data &amp; Analytics, you will act as the technical lead and owner for infrastructure analytics across compute, storage, and networking. You will design and build durable, scalable data pipelines that ingest telemetry from clusters, schedulers, health systems, and capacity trackers into Data Warehouse. You will define and standardize core metrics and semantics (e.g., utilization, occupancy, MFU, goodput, capacity readiness, delivery-to-production). You will architect and maintain self-service dashboards and APIs for fleet, cluster, and squad-level visibility. You will partner closely with stakeholders across Supercomputing Infra, Researchers, Strategy and Executives to ensure metrics reflect operational and business reality.</p>\n<p><strong>Accountabilities</strong></p>\n<ul>\n<li>Act as the technical lead and owner for infrastructure analytics across compute, storage, and networking.</li>\n<li>Design and build durable, scalable data pipelines that ingest telemetry from clusters, schedulers, health systems, and capacity trackers into Data Warehouse.</li>\n</ul>\n<p><strong>The Candidate we&#39;re looking for</strong></p>\n<p><strong>Experience:</strong></p>\n<ul>\n<li>8+ years technical engineering experience with data engineering, analytics, or data science, with increasing technical ownership in startup environment.</li>\n</ul>\n<p><strong>Technical skills:</strong></p>\n<ul>\n<li>Distributed data processing frameworks and large-scale data systems.</li>\n</ul>\n<p><strong>Personal attributes:</strong></p>\n<ul>\n<li>Strong communication skills; can explain complex systems clearly to senior leader.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Software Engineering IC5 – The typical base pay range for this role across the U.S. is USD $139,900 – $274,800 per year.</li>\n<li>Certain roles may be eligible for benefits and other compensation.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2902359a-64d","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Microsoft AI","sameAs":"https://microsoft.ai","logo":"https://logos.yubhub.co/microsoft.ai.png"},"x-apply-url":"https://microsoft.ai/job/member-of-technical-staff-infrastructure-data-analytics-mai-superintelligence-team/","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"USD $139,900 – $274,800 per year","x-skills-required":["data engineering","analytics","data science","distributed data processing frameworks","large-scale data systems"],"x-skills-preferred":["ETL orchestration frameworks","Airflow","Dagster"],"datePosted":"2026-03-06T07:29:22.881Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Multiple Locations, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data engineering, analytics, data science, distributed data processing frameworks, large-scale data systems, ETL orchestration frameworks, Airflow, Dagster","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":139900,"maxValue":274800,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7e974f1a-211"},"title":"Sr Data Engineer","description":"<p>Join our team as a Senior Data Engineer. You&#39;ll develop and maintain data pipelines for our innovative gaming products.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<ul>\n<li>Design, develop, and maintain batch and streaming data pipelines, ensuring seamless data flow and integrity.</li>\n<li>Implement scalable data transformations using dbt and orchestrate workflows via Airflow or equivalent tools.</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>5-7 years of hands-on experience in data engineering.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7e974f1a-211","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Aristocrat","sameAs":"https://aristocrat.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/aristocrat.com.png"},"x-apply-url":"https://aristocrat.wd3.myworkdayjobs.com/en-US/AristocratExternalCareersSite/job/Noida-UP-IN/Sr-Data-Engineer_R0019621","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data engineering","dbt","Airflow"],"x-skills-preferred":["data streaming tools","infrastructure-as-code tools","BI tools"],"datePosted":"2026-03-01T05:05:41.750Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Noida, UP, IN"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data engineering, dbt, Airflow, data streaming tools, infrastructure-as-code tools, BI tools"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_7371adcc-09a"},"title":"Senior Manager Data Engineering","description":"<p>We are seeking a hands-on Engineering Manager to join EA’s Data &amp; Insights – Enterprise Data Engineering (EDE) team. The ideal candidate has extensive experience designing and operating large-scale data platforms, leading complex data engineering initiatives, and working across distributed computing, batch and real-time data processing, and modern lakehouse architectures.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<ul>\n<li>Lead, mentor, and develop a high-performing engineering team through coaching, feedback, and career development.</li>\n<li>Remain hands-on with architecture, design, coding, and technical decisions—contributing directly when needed.</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Bachelor’s or Master’s degree in Computer Science, Engineering, or equivalent degree.</li>\n<li>8+ years of professional experience as a hands-on engineer and technical leader managing multiple projects.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_7371adcc-09a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Electronic Arts","sameAs":"https://jobs.ea.com","logo":"https://logos.yubhub.co/jobs.ea.com.png"},"x-apply-url":"https://jobs.ea.com/en_US/careers/JobDetail/Senior-Manager-Data-Engineering/212290","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Java","SQL","Spark","Kafka","Airflow","Snowflake","AWS","Iceberg"],"x-skills-preferred":["Distributed systems","Modern data technologies","Cloud platforms","Enterprise-scale data environments"],"datePosted":"2026-02-04T13:04:14.005Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hyderabad"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, SQL, Spark, Kafka, Airflow, Snowflake, AWS, Iceberg, Distributed systems, Modern data technologies, Cloud platforms, Enterprise-scale data environments"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4a7597fd-d7a"},"title":"Senior Data Engineer","description":"<p>Joining Razer will place you on a global mission to revolutionize the way the world games. Razer is a place to do great work, offering you the opportunity to make an impact globally while working across a global team located across 5 continents. Razer is also a great place to work, providing you the unique, gamer-centric #LifeAtRazer experience that will put you in an accelerated growth, both personally and professionally.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference. You will collaborate closely with AI/ML engineers, data scientists, and platform teams to build the next generation of AI-driven products.</p>\n<ul>\n<li>Lead AI Data Engineering initiatives by driving the design and development of robust data pipelines for AI/ML workloads, ensuring efficiency, scalability, and reliability.</li>\n<li>Design and implement data architectures that support AI model training, including feature stores, vector databases, and real-time streaming solutions.</li>\n<li>Develop high performance data pipelines that process structured, semi-structured, and unstructured data at scale, supporting the various AI applications</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Hands on experience working with Vector/Graph;Neo4j</li>\n<li>3+ years of experience in data engineering, working on AI/ML-driven data architectures</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4a7597fd-d7a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Razer","sameAs":"https://razer.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/razer.com.png"},"x-apply-url":"https://razer.wd3.myworkdayjobs.com/en-US/Careers/job/Singapore/Senior-Data-Engineer_JR2025005485","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Hands on experience working with Vector/Graph;Neo4j","3+ years of experience in data engineering, working on AI/ML-driven data architectures"],"x-skills-preferred":["Python","SQL","Experience in developing and deploying applications running on cloud infrastructure such as AWS, Azure or Google Cloud Platform using Infrastructure as code tools such as Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes","Experience using orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT)","Excellent with various data processing techniques (both streaming and batch), managing and optimizing data storage (Data Lake, Lake House and Database, SQL, and NoSQL) is essential."],"datePosted":"2026-01-01T15:49:59.491Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Hands on experience working with Vector/Graph;Neo4j, 3+ years of experience in data engineering, working on AI/ML-driven data architectures, Python, SQL, Experience in developing and deploying applications running on cloud infrastructure such as AWS, Azure or Google Cloud Platform using Infrastructure as code tools such as Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes, Experience using orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT), Excellent with various data processing techniques (both streaming and batch), managing and optimizing data storage (Data Lake, Lake House and Database, SQL, and NoSQL) is essential."},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_901a6402-db5"},"title":"Data Engineer","description":"<p>Join Razer to help build and optimize data pipelines and data platforms that support analytics, product improvements, and foundational AI/ML data needs. Collaborate with cross-functional teams to ensure data is reliable, accessible, and governed. Tech stack includes Redshift, Airflow, and DBT.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>Join Razer to help build and optimize data pipelines and data platforms that support analytics, product improvements, and foundational AI/ML data needs. Collaborate with cross-functional teams to ensure data is reliable, accessible, and governed. Tech stack includes Redshift, Airflow, and DBT.</p>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Strong Python and SQL</li>\n<li>Hands-on experience with Redshift, Airflow, DBT</li>\n<li>Mandatory hands-on experience with Apache Spark (batch and/or structured processing)</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_901a6402-db5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Razer","sameAs":"https://razer.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/razer.com.png"},"x-apply-url":"https://razer.wd3.myworkdayjobs.com/en-US/Careers/job/Chengdu/Data-Engineer_JR2025006594","x-work-arrangement":"onsite","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","SQL","Redshift","Airflow","DBT","Apache Spark"],"x-skills-preferred":["Apache Flink","Apache Kafka","Hadoop ecosystem components","ETL design patterns","performance tuning"],"datePosted":"2025-12-26T10:57:30.602Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Chengdu"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, Redshift, Airflow, DBT, Apache Spark, Apache Flink, Apache Kafka, Hadoop ecosystem components, ETL design patterns, performance tuning"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e5eb908e-6f9"},"title":"Senior Data Engineer","description":"<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference.</p>\n<ul>\n<li>Lead AI Data Engineering initiatives by driving the design and development of robust data pipelines for AI/ML workloads, ensuring efficiency, scalability, and reliability.</li>\n<li>Design and implement data architectures that support AI model training, including feature stores, vector databases, and real-time streaming solutions.</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Hands on experience working with Vector/Graph;Neo4j</li>\n<li>3+ years of experience in data engineering, working on AI/ML-driven data architectures</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e5eb908e-6f9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Razer","sameAs":"https://razer.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/razer.com.png"},"x-apply-url":"https://razer.wd3.myworkdayjobs.com/en-US/Careers/job/Singapore/Senior-Data-Engineer_JR2025005485","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Vector/Graph;Neo4j","data engineering","AI/ML-driven data architectures"],"x-skills-preferred":["Python","SQL","Terraform","containerization tools like Dockers","container orchestration platforms like Kubernetes","orchestration tools like Airflow or Prefect","distributed computing framework like Spark or Dask","data transformation tool like Data Build Tool (DBT)"],"datePosted":"2025-12-26T10:53:07.867Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Vector/Graph;Neo4j, data engineering, AI/ML-driven data architectures, Python, SQL, Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes, orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT)"}]}