{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/schema-evolution"},"x-facet":{"type":"skill","slug":"schema-evolution","display":"Schema Evolution","count":3},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_456f029f-2e2"},"title":"Principal Software Engineer","description":"<p>As a Principal Software Engineer on our Go To Market Store (GTM Store) and ZoomInfo Data Platform (ZDP) team, you&#39;ll play a pivotal role in developing ZoomInfo&#39;s next-generation unified data platform.</p>\n<p>You&#39;ll architect and implement infrastructure that powers our GraphQL-based federated query system for seamless data access across platforms including BigTable, BigQuery, and Solr+.</p>\n<p>This is a unique opportunity to influence the technical direction of ZoomInfo&#39;s core data infrastructure, addressing complex challenges such as data freshness, multi-tenant isolation, and real-time data processing at scale.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design and build scalable infrastructure for GTM Store and ZDP with sub-second query latency.</li>\n<li>Architect and implement metadata-driven GraphQL APIs for dynamic schema generation and query federation.</li>\n<li>Develop asynchronous secondary indexing systems for scaling capacity and reducing primary data store load.</li>\n<li>Design real-time analytics streaming data pipelines from BigTable to BigQuery.</li>\n<li>Develop data mutation and deletion frameworks supporting GDPR compliance and schema evolution.</li>\n<li>Implement CDC pipelines and calculated field processing for derived data views.</li>\n<li>Build observability and monitoring solutions for real-time issue diagnosis across distributed data systems.</li>\n<li>Create batch and streaming data processing workflows for complex relationships at scale.</li>\n<li>Collaborate with engineering leaders and product managers to define the technical roadmap.</li>\n<li>Mentor engineers and establish best practices for cloud-native data infrastructure development.</li>\n<li>Partner with cross-functional teams to address data platform requirements and challenges.</li>\n<li>Drive solutions for data freshness, query performance, and system reliability challenges.</li>\n</ul>\n<p>Qualifications:</p>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science, Software Engineering, or related field (or equivalent experience).</li>\n<li>10+ years of software engineering experience building large-scale data platforms.</li>\n<li>Expertise with distributed NoSQL databases and data warehousing systems.</li>\n<li>Strong experience with Java 8+, Scala, Kotlin, GoLang for data systems development.</li>\n<li>Proven experience with GCP or AWS and cloud-native architectures.</li>\n<li>Experience with streaming/real-time data processing technologies.</li>\n<li>Strong system design skills for architecting multi-tenant, distributed systems.</li>\n<li>Hands-on experience with Google Cloud Platform services.</li>\n<li>Knowledge of CDC patterns, event sourcing, and streaming architectures.</li>\n<li>Experience solving data freshness and consistency challenges in distributed systems.</li>\n<li>Background in building observability and monitoring solutions for data platforms.</li>\n<li>Familiarity with metadata management and schema evolution.</li>\n<li>Experience with Kubernetes for deploying data services.</li>\n<li>SQL query optimization and performance tuning expertise.</li>\n<li>Experience building GraphQL APIs with federated or metadata-driven schema generation.</li>\n<li>Strong problem-solving skills and the ability to debug complex distributed systems issues.</li>\n<li>Excellent communication skills for explaining technical decisions to diverse audiences.</li>\n<li>Self-directed with the ability to drive initiatives independently while collaborating with teams.</li>\n<li>Passion for building reliable, observable, and maintainable systems.</li>\n<li>Experience promoting diverse, inclusive work environments.</li>\n</ul>\n<p>Actual compensation offered will be based on factors such as the candidate’s work location, qualifications, skills, experience and/or training. Your recruiter can share more information about the specific salary range for your desired work location during the hiring process.</p>\n<p>We want our employees and their families to thrive. In addition to comprehensive benefits we offer holistic mind, body and lifestyle programs designed for overall well-being. Learn more about ZoomInfo benefits here.</p>\n<p>Below is the US base salary for this position. Additional compensation such as Bonus, Commission, Equity and other benefits may also apply.</p>\n<p>$163,800-$257,400 USD</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_456f029f-2e2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"ZoomInfo","sameAs":"https://www.zoominfo.com/","logo":"https://logos.yubhub.co/zoominfo.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/zoominfo/jobs/8243004002","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$163,800-$257,400 USD","x-skills-required":["Java 8+","Scala","Kotlin","GoLang","GCP","AWS","cloud-native architectures","streaming/real-time data processing technologies","distributed NoSQL databases","data warehousing systems","metadata management","schema evolution","Kubernetes","SQL query optimization","performance tuning","GraphQL APIs","federated or metadata-driven schema generation"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:44:17.604Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Remote-US-CA"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java 8+, Scala, Kotlin, GoLang, GCP, AWS, cloud-native architectures, streaming/real-time data processing technologies, distributed NoSQL databases, data warehousing systems, metadata management, schema evolution, Kubernetes, SQL query optimization, performance tuning, GraphQL APIs, federated or metadata-driven schema generation","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":163800,"maxValue":257400,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_ceba9e5b-250"},"title":"Senior Backend Engineer, Product and Infra","description":"<p>We&#39;re looking for a Senior Backend Engineer to build the systems and services that power our product experience. You&#39;ll own the backend infrastructure that makes our content discoverable, our features responsive, and our platform reliable at scale.</p>\n<p>Your work will directly shape what users experience: designing APIs that serve rich content, building services that handle real-time interactions, implementing content-matching systems for rights and safety, and ensuring our platform performs under load. You&#39;ll architect systems that are fast, correct, and maintainable.</p>\n<p>You&#39;ll collaborate closely with Product, ML Research, and Mobile/Web teams to ship features that matter. We use Python, Go, BigQuery, Pub/Sub, and a microservices architecture,but we care more about good judgment than specific tool experience.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and maintain application-level data models that organize rich content into canonical structures optimized for product features, search, and retrieval.</li>\n<li>Build high-reliability ETLs and streaming pipelines to process usage events, analytics data, behavioral signals, and application logs.</li>\n<li>Develop data services that expose unified content to the application, such as metadata access APIs, indexing workflows, and retrieval-ready representations.</li>\n<li>Implement and refine fingerprinting pipelines used for deduplication, rights attribution, safety checks, and provenance validation.</li>\n<li>Own data consistency between ingestion systems, application surfaces, metadata storage, and downstream reporting environments.</li>\n<li>Define and track key operational metrics, including latency, completeness, accuracy, and event health.</li>\n<li>Collaborate with Product teams to ensure content structures and APIs support evolving features and high-quality user experiences.</li>\n<li>Partner with Analytics and Research teams to deliver clean usage datasets for experimentation, model evaluation, reporting, and internal insights.</li>\n<li>Operate large analytical workloads in BigQuery and build reusable Dataflow/Beam components for structured processing.</li>\n<li>Improve reliability and scale by designing robust schema evolution strategies, idempotent pipelines, and well-instrumented operational flows.</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Experience building production backend services and APIs at scale</li>\n<li>Experience building ETL/ELT pipelines, event processing systems, and structured data models for applications or analytics</li>\n<li>Strong background in data modeling, metadata systems, indexing, or building canonical representations for heterogeneous content</li>\n<li>Proficiency in Python, Go, SQL, and scalable data-processing frameworks (Dataflow/Beam, Spark, or similar)</li>\n<li>Familiarity with BigQuery or other analytical data warehouses and strong comfort optimizing large queries and schemas</li>\n<li>Experience with event-driven architectures, Pub/Sub, or Kafka-like systems</li>\n<li>Strong understanding of data quality, schema evolution, lineage, and operational reliability</li>\n<li>Ability to design pipelines that balance cost, latency, correctness, and scale</li>\n<li>Clear communication skills and an ability to collaborate closely with Product, Research, and Analytics stakeholders</li>\n</ul>\n<p><strong>Nice to Have</strong></p>\n<ul>\n<li>Experience building application-facing APIs or microservices that expose structured content</li>\n<li>Background in information retrieval, indexing systems, or search infrastructure</li>\n<li>Experience with fingerprinting, perceptual hashing, audio similarity metrics, or content-matching algorithms</li>\n<li>Familiarity with ML workflows and how downstream analytics and usage data feed back into research pipelines</li>\n<li>Understanding of batch + streaming architectures and how to blend them effectively</li>\n<li>Experience with Go, Next.js, or React Native for occasional full-stack contributions</li>\n</ul>\n<p><strong>Why Join Us</strong></p>\n<p>You will design the core data services and pipelines that power our product experience, analytics, and business operations. You’ll work on high-impact data challenges involving real-time signals, large-scale metadata systems, and cross-platform consistency. You’ll join a small, fast-moving team where you’ll shape the structure, reliability, and intelligence of our downstream data ecosystem.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Highly competitive salary and equity</li>\n<li>Quarterly productivity budget</li>\n<li>Flexible time off</li>\n<li>Fantastic office location in Manhattan</li>\n<li>Productivity package, including ChatGPT Plus, Claude Code, and Copilot</li>\n<li>Top-notch private health, dental, and vision insurance for you and your dependents</li>\n<li>401(k) plan options with employer matching</li>\n<li>Concierge medical/primary care through One Medical and Rightway</li>\n<li>Mental health support from Spring Health</li>\n<li>Personalized life insurance, travel assistance, and many other perks</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_ceba9e5b-250","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Udio","sameAs":"https://www.udio.com/","logo":"https://logos.yubhub.co/udio.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/udio/jobs/4987729008","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$180,000 - $220,000","x-skills-required":["Python","Go","BigQuery","Pub/Sub","Data modeling","Metadata systems","Indexing","Canonical representations","ETL/ELT pipelines","Event processing systems","Structured data models","Scalable data-processing frameworks","Analytical data warehouses","Event-driven architectures","Kafka-like systems","Data quality","Schema evolution","Lineage","Operational reliability"],"x-skills-preferred":["Application-facing APIs","Microservices","Information retrieval","Indexing systems","Search infrastructure","Fingerprinting","Perceptual hashing","Audio similarity metrics","Content-matching algorithms","ML workflows","Batch + streaming architectures"],"datePosted":"2026-04-17T13:05:20.076Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Go, BigQuery, Pub/Sub, Data modeling, Metadata systems, Indexing, Canonical representations, ETL/ELT pipelines, Event processing systems, Structured data models, Scalable data-processing frameworks, Analytical data warehouses, Event-driven architectures, Kafka-like systems, Data quality, Schema evolution, Lineage, Operational reliability, Application-facing APIs, Microservices, Information retrieval, Indexing systems, Search infrastructure, Fingerprinting, Perceptual hashing, Audio similarity metrics, Content-matching algorithms, ML workflows, Batch + streaming architectures","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_be821069-a7f"},"title":"Asset Data Engineer","description":"<p>Join the Asset Data team and build the streaming data infrastructure that powers Anchorage&#39;s digital asset platform. You&#39;ll design systems that ingest real-time blockchain and market data from diverse providers, transforming raw feeds into certified, trusted data products.</p>\n<p>We&#39;re creating contract-governed supply chains that let us onboard new assets and providers quickly while maintaining the low-latency, high-availability SLOs our business depends on.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Build streaming data pipelines for blockchain data (onchain transactions, staking rewards, validator info) and market data (prices, trades, order books)</li>\n<li>Design and implement data contracts and validation gates that enforce quality and schema compliance at ingestion points</li>\n</ul>\n<p><strong>Complexity and Impact of Work:</strong></p>\n<ul>\n<li>Collaborate on designing the architecture for standardized ingestion patterns that enable rapid onboarding of new blockchains and market data feeds</li>\n<li>Establish redundancy and failover patterns to meet Tier 1 availability and freshness SLOs for critical data products</li>\n</ul>\n<p><strong>Organizational Knowledge:</strong></p>\n<ul>\n<li>Collaborate with Protocols, Trading, and Custody teams to understand their data needs and design certified data products with clear SLAs</li>\n<li>Partner with Data Platform team on orchestration, storage patterns (BigLake), and metadata management (Atlan)</li>\n</ul>\n<p><strong>Communication and Influence:</strong></p>\n<ul>\n<li>Advocate for contract-governed data supply chains and help establish engineering standards for producer patterns across the org</li>\n<li>Contribute to architectural decisions and help mature the team&#39;s practices around observability, testing, and operational excellence</li>\n</ul>\n<p><strong>Requirements:</strong></p>\n<ul>\n<li>5-7+ years building streaming or high-throughput data systems: You have experience designing and operating production data pipelines that handle large volumes with low latency and high reliability</li>\n<li>Solid backend engineering skills: You&#39;re proficient in Go or Python and have built services that interact with streaming infrastructure (Kafka, pub/sub, websockets, REST APIs)</li>\n<li>Blockchain data familiarity: You understand blockchain concepts and are comfortable working with on-chain data (transactions, events, staking, validators) across multiple chains with different data models</li>\n<li>Data engineering adjacent skills: You&#39;re comfortable with data transformation patterns, schema evolution, and working with cloud data warehouses (BigQuery) and storage systems (GCS, BigLake)</li>\n<li>Operational mindset: You have experience deploying and operating services on cloud platforms (preferably GCP), with strong practices around monitoring, alerting, and incident response</li>\n</ul>\n<p><strong>Preferred Qualifications:</strong></p>\n<ul>\n<li>Staking data expertise: You&#39;ve worked with staking rewards, validator data, or proof-of-stake blockchain infrastructure</li>\n<li>Market data systems: You&#39;ve built systems that ingest and process market data (prices, trades, order books) from exchanges or data vendors</li>\n<li>Infrastructure as code: You have experience with Terraform, Kubernetes, and modern DevOps practices</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_be821069-a7f","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anchorage Digital","sameAs":"https://www.anchorage.co/","logo":"https://logos.yubhub.co/anchorage.co.png"},"x-apply-url":"https://jobs.lever.co/anchorage/82139746-fb0e-44b9-bbb6-ae078e5d251a","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Go","Python","Kafka","pub/sub","websockets","REST APIs","blockchain data","data transformation patterns","schema evolution","cloud data warehouses","storage systems"],"x-skills-preferred":["stake data expertise","market data systems","infrastructure as code"],"datePosted":"2026-04-17T12:25:02.799Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Go, Python, Kafka, pub/sub, websockets, REST APIs, blockchain data, data transformation patterns, schema evolution, cloud data warehouses, storage systems, stake data expertise, market data systems, infrastructure as code"}]}