{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/data-processing-technologies"},"x-facet":{"type":"skill","slug":"data-processing-technologies","display":"Data Processing Technologies","count":3},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ba30b234-c68"},"title":"Senior Data Engineer, Payments","description":"<p>We&#39;re looking for a Senior Data Engineer to join our Payments team. As a critical part of our operations, you&#39;ll handle data related to compliance with Tax, Payments, and Legal regulations. You&#39;ll design, build, and maintain robust and efficient data pipelines that collect, process, and store data from various sources, including user interactions, listing details, and external data feeds.</p>\n<p>Your work will involve developing data models that enable the efficient analysis and manipulation of data for merchandising optimization, ensuring data quality, consistency, and accuracy. You&#39;ll also develop high-quality data assets for product use-cases by partnering with Product, AI/ML, and Data Science teams.</p>\n<p>As a Senior Data Engineer, you&#39;ll contribute to creating standards and best practices for Airbnb&#39;s Data Engineering and shape the tools, processes, and standards used by the broader data community. You&#39;ll collaborate with cross-functional teams to define data requirements and deliver data solutions that drive merchandising and sales improvements.</p>\n<p>To succeed in this role, you&#39;ll need 6+ years of relevant industry experience, a BE/B.Tech in Computer Science or a relevant technical degree, and hands-on experience in DSA coding, data structure, and algorithm. You&#39;ll also need extensive experience designing, building, and operating robust distributed data platforms and handling data at the petabyte scale.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ba30b234-c68","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Airbnb","sameAs":"https://www.airbnb.com/","logo":"https://logos.yubhub.co/airbnb.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/airbnb/jobs/7256787","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Scala","Python","data processing technologies","query authoring (SQL)","ETL schedulers (Apache Airflow, Luigi, Oozie, AWS Glue)","data warehousing concepts","relational databases (PostgreSQL, MySQL)","columnar databases (Redshift, BigQuery, HBase, ClickHouse)"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:13.348Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bangalore, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Scala, Python, data processing technologies, query authoring (SQL), ETL schedulers (Apache Airflow, Luigi, Oozie, AWS Glue), data warehousing concepts, relational databases (PostgreSQL, MySQL), columnar databases (Redshift, BigQuery, HBase, ClickHouse)"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_456f029f-2e2"},"title":"Principal Software Engineer","description":"<p>As a Principal Software Engineer on our Go To Market Store (GTM Store) and ZoomInfo Data Platform (ZDP) team, you&#39;ll play a pivotal role in developing ZoomInfo&#39;s next-generation unified data platform.</p>\n<p>You&#39;ll architect and implement infrastructure that powers our GraphQL-based federated query system for seamless data access across platforms including BigTable, BigQuery, and Solr+.</p>\n<p>This is a unique opportunity to influence the technical direction of ZoomInfo&#39;s core data infrastructure, addressing complex challenges such as data freshness, multi-tenant isolation, and real-time data processing at scale.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build scalable infrastructure for GTM Store and ZDP with sub-second query latency.</li>\n<li>Architect and implement metadata-driven GraphQL APIs for dynamic schema generation and query federation.</li>\n<li>Develop asynchronous secondary indexing systems for scaling capacity and reducing primary data store load.</li>\n<li>Design real-time analytics streaming data pipelines from BigTable to BigQuery.</li>\n<li>Develop data mutation and deletion frameworks supporting GDPR compliance and schema evolution.</li>\n<li>Implement CDC pipelines and calculated field processing for derived data views.</li>\n<li>Build observability and monitoring solutions for real-time issue diagnosis across distributed data systems.</li>\n<li>Create batch and streaming data processing workflows for complex relationships at scale.</li>\n<li>Collaborate with engineering leaders and product managers to define the technical roadmap.</li>\n<li>Mentor engineers and establish best practices for cloud-native data infrastructure development.</li>\n<li>Partner with cross-functional teams to address data platform requirements and challenges.</li>\n<li>Drive solutions for data freshness, query performance, and system reliability challenges.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, Software Engineering, or related field (or equivalent experience).</li>\n<li>10+ years of software engineering experience building large-scale data platforms.</li>\n<li>Expertise with distributed NoSQL databases and data warehousing systems.</li>\n<li>Strong experience with Java 8+, Scala, Kotlin, GoLang for data systems development.</li>\n<li>Proven experience with GCP or AWS and cloud-native architectures.</li>\n<li>Experience with streaming/real-time data processing technologies.</li>\n<li>Strong system design skills for architecting multi-tenant, distributed systems.</li>\n<li>Hands-on experience with Google Cloud Platform services.</li>\n<li>Knowledge of CDC patterns, event sourcing, and streaming architectures.</li>\n<li>Experience solving data freshness and consistency challenges in distributed systems.</li>\n<li>Background in building observability and monitoring solutions for data platforms.</li>\n<li>Familiarity with metadata management and schema evolution.</li>\n<li>Experience with Kubernetes for deploying data services.</li>\n<li>SQL query optimization and performance tuning expertise.</li>\n<li>Experience building GraphQL APIs with federated or metadata-driven schema generation.</li>\n<li>Strong problem-solving skills and the ability to debug complex distributed systems issues.</li>\n<li>Excellent communication skills for explaining technical decisions to diverse audiences.</li>\n<li>Self-directed with the ability to drive initiatives independently while collaborating with teams.</li>\n<li>Passion for building reliable, observable, and maintainable systems.</li>\n<li>Experience promoting diverse, inclusive work environments.</li>\n</ul>\n<p>Actual compensation offered will be based on factors such as the candidate’s work location, qualifications, skills, experience and/or training. Your recruiter can share more information about the specific salary range for your desired work location during the hiring process.</p>\n<p>We want our employees and their families to thrive. In addition to comprehensive benefits we offer holistic mind, body and lifestyle programs designed for overall well-being. Learn more about ZoomInfo benefits here.</p>\n<p>Below is the US base salary for this position. Additional compensation such as Bonus, Commission, Equity and other benefits may also apply.</p>\n<p>$163,800-$257,400 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_456f029f-2e2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8243004002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$163,800-$257,400 USD","x-skills-required":["Java 8+","Scala","Kotlin","GoLang","GCP","AWS","cloud-native architectures","streaming/real-time data processing technologies","distributed NoSQL databases","data warehousing systems","metadata management","schema evolution","Kubernetes","SQL query optimization","performance tuning","GraphQL APIs","federated or metadata-driven schema generation"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:17.604Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-US-CA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java 8+, Scala, Kotlin, GoLang, GCP, AWS, cloud-native architectures, streaming/real-time data processing technologies, distributed NoSQL databases, data warehousing systems, metadata management, schema evolution, Kubernetes, SQL query optimization, performance tuning, GraphQL APIs, federated or metadata-driven schema generation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":163800,"maxValue":257400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_448a56f3-ab5"},"title":"Director of Data Engineering and Agentic AI Automation, Finance","description":"<p><strong>Director of Data Engineering and Agentic AI Automation, Finance</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Department</strong></p>\n<p>Finance</p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$347K – $490K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>We are looking for a Director of Data Engineering and Agentic AI Automation to lead the next generation of our finance data infrastructure. As OpenAI expands its Finance operations, we need scalable and trustworthy data systems to match the pace and complexity of our growth. This includes well-modeled, auditable data for revenue recognition, financial reporting, and planning, supported by reliable pipelines that connect ERP, planning, and operational systems. You will lead a group of analytics engineers, data engineers, and AI engineers to build the data pipelines that connect our internal engineering systems with enterprise platforms such as Oracle Fusion ERP. This role will also define the roadmap for agentic AI automation, enabling intelligent workflows, process automation, and AI-driven decision-making across Finance.</p>\n<p><strong>In this role, you will:</strong></p>\n<ul>\n<li>Build and maintain scalable, auditable data infrastructure that powers accurate financial information, with a focus on revenue recognition, compute attribution, and close automation.</li>\n</ul>\n<ul>\n<li>Lead and grow teams of analytics engineers, data engineers, and AI engineers to deliver high-impact, intelligent data systems.</li>\n</ul>\n<ul>\n<li>Guide work across financial close and allocations automation, B2C revenue automation from engineering systems to ERP (including reconciliation with cash and source systems), and other mission-critical financial processes.</li>\n</ul>\n<ul>\n<li>Design and implement data pipelines connecting ERP, planning, and operational systems, including Oracle Fusion, Anaplan, and Workday.</li>\n</ul>\n<ul>\n<li>Build and support scalable, audit-proof architecture that enables reliable financial reporting and compliance.</li>\n</ul>\n<ul>\n<li>Develop data and AI-powered workflows that enhance forecasting accuracy, compliance automation, and operational efficiency.</li>\n</ul>\n<ul>\n<li>Create and maintain data marts and products that support stakeholders across Revenue, FP&amp;A, Tax, Procurement, Hardware Accounting, and Controller teams.</li>\n</ul>\n<ul>\n<li>Define and enforce best practices for data modeling, lineage, observability, and reconciliation across finance data domains.</li>\n</ul>\n<ul>\n<li>Set the technical direction and manage team structure, mentoring engineers and overseeing contractors or system integrators to ensure delivery of high-quality outcomes.</li>\n</ul>\n<ul>\n<li>Partner with senior leaders across Finance, Engineering, and Infrastructure to align on priorities and integrate new automation capabilities.</li>\n</ul>\n<ul>\n<li>Ensure data systems are AI-ready and capable of supporting predictive analytics, autonomous agent workflows, and large-scale automation.</li>\n</ul>\n<ul>\n<li>Own and maintain Tier-1 data pipelines with strict SLA, data quality, and compliance standards.</li>\n</ul>\n<ul>\n<li>Drive the long-term roadmap for agentic AI enablement to build the foundation for “Finance on OpenAI.”</li>\n</ul>\n<p><strong>You might thrive in this role if you have:</strong></p>\n<ul>\n<li>12+ years in data engineering, with proven experience building and managing enterprise-scale, auditable ETL pipelines and complex datasets</li>\n</ul>\n<ul>\n<li>Proficiency in SQL and Python, with demonstrated experience in schema design, data modeling, and orchestration frameworks</li>\n</ul>\n<ul>\n<li>Expertise in distributed data processing technologies such as Apache Spark, Kafka, and cloud-native storage (e.g., S3, ADLS)</li>\n</ul>\n<ul>\n<li>Deep knowledge of enterprise data architecture, especially within Finance and Supply Chain</li>\n</ul>\n<ul>\n<li>Familiarity with financial processes (close, allocations, revenue recognition) and supply chain data models (Supply and demand planning, procurement, vendor master), along with experience in ingesting data from internal engineering systems with large volumes of B2C</li>\n</ul>\n<ul>\n<li>Experience integrating with contract manufacturers and external logistics providers is a strong plus</li>\n</ul>\n<ul>\n<li>Strong track record of partnering with senior business stakeholders</li>\n</ul>\n<p><strong>Work Environment</strong></p>\n<p>This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_448a56f3-ab5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/e84e7b7e-a82e-411e-929a-615dc3080280","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$347K – $490K • Offers Equity","x-skills-required":["SQL","Python","Apache Spark","Kafka","cloud-native storage","data modeling","orchestration frameworks","distributed data processing technologies","enterprise data architecture","financial processes","supply chain data models"],"x-skills-preferred":["ETL pipelines","complex datasets","schema design","data engineering","data infrastructure","auditable data","revenue recognition","financial reporting","planning","ERP","planning","operational systems","Oracle Fusion","Anaplan","Workday","data marts","products","stakeholders","Revenue","FP&A","Tax","Procurement","Hardware Accounting","Controller","data modeling","lineage","observability","reconciliation","finance data domains","team structure","engineers","contractors","system integrators","predictive analytics","autonomous agent workflows","large-scale automation"],"datePosted":"2026-03-06T18:27:50.931Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Apache Spark, Kafka, cloud-native storage, data modeling, orchestration frameworks, distributed data processing technologies, enterprise data architecture, financial processes, supply chain data models, ETL pipelines, complex datasets, schema design, data engineering, data infrastructure, auditable data, revenue recognition, financial reporting, planning, ERP, planning, operational systems, Oracle Fusion, Anaplan, Workday, data marts, products, stakeholders, Revenue, FP&A, Tax, Procurement, Hardware Accounting, Controller, data modeling, lineage, observability, reconciliation, finance data domains, team structure, engineers, contractors, system integrators, predictive analytics, autonomous agent workflows, large-scale automation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":347000,"maxValue":490000,"unitText":"YEAR"}}}]}