{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/trino"},"x-facet":{"type":"skill","slug":"trino","display":"Trino","count":9},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_22bcbb50-ef4"},"title":"Member of Technical Staff - Data Platform","description":"<p><strong>About the Role</strong></p>\n<p>The Data Platform team at xAI builds and operates the infrastructure responsible for all large-scale data transport and processing across the company.</p>\n<p>As a software engineer on the Data Platform team, you will design, build, and operate the distributed systems powering X&#39;s data movement and compute.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and implement high-throughput, low-latency data ingestion and transport systems.</li>\n<li>Scale and optimise multi-tenant Kafka infrastructure supporting real-time workloads.</li>\n<li>Extend and tune Spark, Flink, and Trino for demanding production pipelines.</li>\n<li>Build interfaces, APIs, and pipelines enabling teams to query, process, and move data at petabyte scale.</li>\n<li>Debug and optimise distributed systems, with a focus on reliability and performance under load.</li>\n<li>Collaborate with ML, product, and infrastructure teams to unblock critical data workflows.</li>\n</ul>\n<p><strong>Basic Qualifications</strong></p>\n<ul>\n<li>Proven expertise in distributed systems, stream processing, or large-scale data platforms.</li>\n<li>Proficiency in Rust, Go, Scala or similar systems languages.</li>\n<li>Hands-on experience with Kafka, Flink, Spark, Trino, or Hadoop in production.</li>\n<li>Strong debugging, profiling, and performance optimisation skills.</li>\n<li>Track record of shipping and maintaining critical infrastructure.</li>\n<li>Comfortable working in fast-moving, high-stakes environments with minimal guardrails.</li>\n</ul>\n<p><strong>Compensation and Benefits</strong></p>\n<p>$180,000 - $440,000 USD</p>\n<p>Base salary is just one part of our total rewards package at X, which also includes equity, comprehensive medical, vision, and dental coverage, access to a 401(k) retirement plan, short &amp; long-term disability insurance, life insurance, and various other discounts and perks.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_22bcbb50-ef4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.x.com/","logo":"https://logos.yubhub.co/x.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/4803862007","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$180,000 - $440,000 USD","x-skills-required":["Rust","Go","Scala","Kafka","Flink","Spark","Trino","Hadoop","distributed systems","stream processing","large-scale data platforms"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:57:30.705Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Palo Alto, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Rust, Go, Scala, Kafka, Flink, Spark, Trino, Hadoop, distributed systems, stream processing, large-scale data platforms","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":440000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_40bd1c3c-ae0"},"title":"Software Engineer - Trino Engine","description":"<p>We&#39;re looking for strong Java engineers to work with our globally distributed engineering team on the core of Starburst&#39;s software. This role will allow you to deepen your expertise in a rapidly evolving technology and make a significant impact on leading data analytics products.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Design, develop and maintain core components in Open Source Trino, the Starburst Enterprise Platform, or the Starburst Galaxy</li>\n<li>Research and Improve performance of Trino query engine on complex queries without sacrificing correctness of the results</li>\n<li>Collaborate with your team members and other teams globally and operate in a fast-paced environment</li>\n<li>We prioritize focused work, ensuring minimal time is spent in formal meetings, allowing you to concentrate on coding and PR reviews</li>\n<li>Being able to clearly articulate your ideas in writing across various communication channels like Slack, GitHub PRs, and Design Documents is essential in our globally distributed team</li>\n<li>Provide exceptional customer support for both internal and external customers</li>\n</ul>\n<p>Some of the things we look for:</p>\n<ul>\n<li>At least 2 years of experience developing distributed systems</li>\n<li>Software development experience with Java</li>\n<li>Demonstrated experience with software engineering and design best practices</li>\n<li>Appreciation for creating maintainable, performant, and high-quality software as part of a fun, high-performing global team</li>\n<li>Interest in distributed systems or database internals such as query optimization</li>\n<li>Intrinsic motivation for improving your software engineering craftsmanship</li>\n<li>Demonstration of ownership, grit, and bias for action - core values at Starburst.</li>\n</ul>\n<p>Bonus points:</p>\n<ul>\n<li>Prior experience with database internals such as query optimization</li>\n<li>Familiarity with Trino</li>\n<li>Experience in contributing to larger scale Open-Source Software</li>\n</ul>\n<p>Where could this role be based?</p>\n<ul>\n<li>This role is based in our Warsaw office and follows a hybrid model, with an expectation of being onsite 1-2 days per week.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_40bd1c3c-ae0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Starburst","sameAs":"https://www.starburst.io/","logo":"https://logos.yubhub.co/starburst.io.png"},"x-apply-url":"https://job-boards.greenhouse.io/starburst/jobs/4783675008","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":"310 000 zł-403 000 zł PLN","x-skills-required":["Java","Distributed systems","Software engineering","Database internals","Query optimization"],"x-skills-preferred":["Trino","Open-Source Software"],"datePosted":"2026-04-18T15:56:46.902Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Warsaw, Poland"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Java, Distributed systems, Software engineering, Database internals, Query optimization, Trino, Open-Source Software"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_059293a1-afa"},"title":"Systems Engineer, Data","description":"<p>About Us</p>\n<p>At Cloudflare, we are on a mission to help build a better Internet. Today the company runs one of the world’s largest networks that powers millions of websites and other Internet properties for customers ranging from individual bloggers to SMBs to Fortune 500 companies.</p>\n<p>We protect and accelerate any Internet application online without adding hardware, installing software, or changing a line of code. Internet properties powered by Cloudflare all have web traffic routed through its intelligent global network, which gets smarter with every request. As a result, they see significant improvement in performance and a decrease in spam and other attacks.</p>\n<p>We were named to Entrepreneur Magazine’s Top Company Cultures list and ranked among the World’s Most Innovative Companies by Fast Company.</p>\n<p>About the Team</p>\n<p>The Core Data team’s mission is building a centralized data platform for Cloudflare that provides secure, democratized access to data for internal customers throughout the company. We operate infrastructure and craft tools to empower both technical and non-technical users to answer their most important questions. We facilitate access to data from federated sources across the company for dashboarding, ad-hoc querying and in-product use cases. We power data pipelines and data products, secure and monitor data, and drive data governance at Cloudflare.</p>\n<p>Our work enables every individual at the company to act with greater information and make more informed decisions.</p>\n<p>About the Role</p>\n<p>We are looking for a systems engineer with a strong background in data to help us expand and maintain our data infrastructure. You’ll contribute to the technical implementation of our scaling data platform, manage access while accounting for privacy and security, build data pipelines, and develop tools to automate accessibility and usefulness of data. You’ll collaborate with teams including Product Growth, Marketing, and Billing to help them make informed decisions and power usage-based invoicing platforms, as well as work with product teams to bring new data-driven solutions to Cloudflare customers.</p>\n<p>Responsibilities</p>\n<ul>\n<li>Contribute to the design and execution of technical architecture for highly visible data infrastructure at the company.</li>\n<li>Design and develop tools and infrastructure to improve and scale our data systems at Cloudflare.</li>\n<li>Build and maintain data pipelines and data products to serve customers throughout the company, including tools to automate delivery of those services.</li>\n<li>Gain deep knowledge of our data platforms and tools to guide and enable stakeholders with their data needs.</li>\n<li>Work across our tech stack, which includes Kubernetes, Trino, Iceberg, Clickhouse, and PostgreSQL, with software built using Go, Javascript/Typescript, Python, and others.</li>\n<li>Collaborate with peers to reinforce a culture of exceptional delivery and accountability on the team.</li>\n</ul>\n<p>Requirements</p>\n<ul>\n<li>3-5+ years of experience as a software engineer with a focus on building and maintaining data infrastructure.</li>\n<li>Experience participating in technical initiatives in a cross-functional context, working with stakeholders to deliver value.</li>\n<li>Practical experience with data infrastructure components, such as Trino, Spark, Iceberg/Delta Lake, Kafka, Clickhouse, or PostgreSQL.</li>\n<li>Hands-on experience building and debugging data pipelines.</li>\n<li>Proficient using backend languages like Go, Python, or Typescript, along with strong SQL skills.</li>\n<li>Strong analytical skills, with a focus on understanding how data is used to drive business value.</li>\n<li>Solid communication skills, with the ability to explain technical concepts to both technical and non-technical audiences.</li>\n</ul>\n<p>Desirable Skills</p>\n<ul>\n<li>Experience with data orchestration and infrastructure platforms like Airflow and DBT.</li>\n<li>Experience deploying and managing services in Kubernetes.</li>\n<li>Familiarity with data governance processes, privacy requirements, or auditability.</li>\n<li>Interest in or knowledge of machine learning models and MLOps.</li>\n</ul>\n<p>What Makes Cloudflare Special?</p>\n<p>We’re not just a highly ambitious, large-scale technology company. We’re a highly ambitious, large-scale technology company with a soul. Fundamental to our mission to help build a better Internet is protecting the free and open Internet.</p>\n<p>Project Galileo: Since 2014, we&#39;ve equipped more than 2,400 journalism and civil society organizations in 111 countries with powerful tools to defend themselves against attacks that would otherwise censor their work, technology already used by Cloudflare’s enterprise customers--at no cost.</p>\n<p>Athenian Project: In 2017, we created the Athenian Project to ensure that state and local governments have the highest level of protection and reliability for free, so that their constituents have access to election information and voter registration. Since the project, we&#39;ve provided services to more than 425 local government election websites in 33 states.</p>\n<p>1.1.1.1: We released 1.1.1.1 to help fix the foundation of the Internet by building a faster, more secure and privacy-centric public DNS resolver. This is available publicly for everyone to use - it is the first consumer-focused service Cloudflare has ever released.</p>\n<p>Here’s the deal - we don’t store client IP addresses never, ever. We will continue to abide by our privacy commitment and ensure that no user data is sold to advertisers or used to target consumers.</p>\n<p>Sound like something you’d like to be a part of? We’d love to hear from you!</p>\n<p>This position may require access to information protected under U.S. export control laws, including the U.S. Export Administration Regulations. Please note that any offer of employment may be conditioned on your authorization to receive software or technology controlled under these U.S. export laws without sponsorship for an export license.</p>\n<p>Cloudflare is proud to be an equal opportunity employer. We are committed to providing equal employment opportunity for all people and place great value in both diversity and inclusiveness. All qualified applicants will be considered for employment without regard to their, or any other person&#39;s, perceived or actual race, color, religion, sex, gender, gender identity, gender expression, sexual orientation, national origin, ancestry, citizenship, age, physical or mental disability, medical condition, family care status, or any other basis protected by law. We are an AA/Veterans/Disabled Employer. Cloudflare provides reasonable accommodations to qualified individuals with disabilities. Please tell us if you require a reasonable accommodation to apply for a job. Examples of reasonable accommodations include, but are not limited to, changing the application process, providing documents in an alternate format, using a sign language interpreter, or using specialized equipment. If you require a reasonable accommodation to apply for a job, please contact us via e-mail at hr@cloudflare.com or via mail at 101 Townsend St. San Francisco, CA 94107.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_059293a1-afa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Cloudflare","sameAs":"https://www.cloudflare.com/","logo":"https://logos.yubhub.co/cloudflare.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/cloudflare/jobs/7527453","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data infrastructure","data pipelines","data products","Kubernetes","Trino","Iceberg","Clickhouse","PostgreSQL","Go","Javascript/Typescript","Python","SQL"],"x-skills-preferred":["data orchestration","infrastructure platforms","Airflow","DBT","machine learning models","MLOps"],"datePosted":"2026-04-18T15:50:12.541Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Hybrid"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"data infrastructure, data pipelines, data products, Kubernetes, Trino, Iceberg, Clickhouse, PostgreSQL, Go, Javascript/Typescript, Python, SQL, data orchestration, infrastructure platforms, Airflow, DBT, machine learning models, MLOps"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c6d7f1a0-882"},"title":"Resident Solutions Architect - Mumbai","description":"<p>We are seeking an experienced Resident Solution Architect (RSA) to join our Professional Services team and work directly with strategic customers on their data and AI transformation initiatives using the Databricks platform.</p>\n<p>As an RSA, you will serve as a trusted technical advisor and hands-on expert, guiding customers to solve complex big data challenges using the Databricks platform.</p>\n<p>Key responsibilities include:</p>\n<ul>\n<li>Collaborating with customers to understand their data and AI transformation goals and developing tailored solutions using the Databricks platform</li>\n<li>Designing and implementing scalable and secure data architectures using Apache Spark, Delta Lake, and other Databricks technologies</li>\n<li>Providing expert-level technical guidance and support to customers during the implementation process</li>\n<li>Identifying and addressing potential roadblocks and providing creative solutions to overcome them</li>\n</ul>\n<p>Requirements include:</p>\n<ul>\n<li>10+ years of experience with Big Data Technologies such as Apache Spark, Kafka, and Data Lakes in a customer-facing post-sales, technical architecture, or consulting role</li>\n<li>4+ years of experience as a Solution Architect creating designs, solving Big Data challenges for customers</li>\n<li>Expertise in Apache Spark, distributed computing, and Databricks platform capabilities</li>\n<li>Comfortable writing code in Python, PySpark, and Scala</li>\n<li>Exceptional SQL, Spark SQL, Spark-streaming skills</li>\n<li>Advanced knowledge of Spark optimizations, Delta, Databricks Lakehouse Platforms</li>\n<li>Expertise in Azure</li>\n<li>Expertise in NoSQL databases (MongoDB, Redis, HBase)</li>\n<li>Expertise in data governance and security (Unity Catalog, RBAC)</li>\n<li>Ability to work with Partner Organization and deliver complex programs</li>\n<li>Ability to lead large technical delivery teams</li>\n<li>Understands the larger competitive landscape, such as EMR, Snowflake, and Sagemaker</li>\n<li>Experience of migration from On-prem / Cloud to Databricks is a plus</li>\n<li>Excellent communication and client-facing consulting skills, with the ability to simplify complex technical concepts</li>\n<li>Willingness to travel for onsite customer engagements within India</li>\n<li>Documentation and white-boarding skills</li>\n</ul>\n<p>Good-to-have Skills:</p>\n<ul>\n<li>Experience with ML libraries/frameworks: Scikit-learn, TensorFlow, PyTorch</li>\n<li>Familiarity with MLOps tools and processes, including MLflow for tracking and deployment</li>\n<li>Experience delivering LLM and GenAI solutions at scale (RAG architectures, prompt engineering)</li>\n<li>Extensive experience on Hadoop, Trino, Ranger and other open-source technology stack</li>\n<li>Expertise on cloud platforms like AWS and GCP</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c6d7f1a0-882","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Databricks","sameAs":"https://databricks.com","logo":"https://logos.yubhub.co/databricks.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/databricks/jobs/8107166002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Apache Spark","Kafka","Data Lakes","Python","PySpark","Scala","SQL","Spark SQL","Spark-streaming","Azure","NoSQL databases","data governance","security","Unity Catalog","RBAC"],"x-skills-preferred":["ML libraries/frameworks","MLOps tools and processes","LLM and GenAI solutions","Hadoop","Trino","Ranger","AWS","GCP"],"datePosted":"2026-04-18T15:45:04.317Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mumbai, India"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Apache Spark, Kafka, Data Lakes, Python, PySpark, Scala, SQL, Spark SQL, Spark-streaming, Azure, NoSQL databases, data governance, security, Unity Catalog, RBAC, ML libraries/frameworks, MLOps tools and processes, LLM and GenAI solutions, Hadoop, Trino, Ranger, AWS, GCP"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_9b657c4e-8a1"},"title":"Member of Technical Staff - Data Platform","description":"<p><strong>About the Role</strong></p>\n<p>As a software engineer on the Data Platform team, you will design, build, and operate the distributed systems powering X&#39;s data movement and compute. You will take ownership of infrastructure components that process trillions of events daily, driving the scalability, performance, and reliability of the systems that power product and ML workloads across the company.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Design and implement high-throughput, low-latency data ingestion and transport systems.</li>\n<li>Scale and optimize multi-tenant Kafka infrastructure supporting real-time workloads.</li>\n<li>Extend and tune Spark, Flink, and Trino for demanding production pipelines.</li>\n<li>Build interfaces, APIs, and pipelines enabling teams to query, process, and move data at petabyte scale.</li>\n<li>Debug and optimize distributed systems, with a focus on reliability and performance under load.</li>\n<li>Collaborate with ML, product, and infrastructure teams to unblock critical data workflows.</li>\n</ul>\n<p><strong>Basic Qualifications</strong></p>\n<ul>\n<li>Proven expertise in distributed systems, stream processing, or large-scale data platforms.</li>\n<li>Proficiency in Rust, Go, Scala or similar systems languages.</li>\n<li>Hands-on experience with Kafka, Flink, Spark, Trino, or Hadoop in production.</li>\n<li>Strong debugging, profiling, and performance optimization skills.</li>\n<li>Track record of shipping and maintaining critical infrastructure.</li>\n<li>Comfortable working in fast-moving, high-stakes environments with minimal guardrails.</li>\n</ul>\n<p><strong>Compensation and Benefits</strong></p>\n<p>$180,000 - $440,000 USD</p>\n<p>Base salary is just one part of our total rewards package at X, which also includes equity, comprehensive medical, vision, and dental coverage, access to a 401(k) retirement plan, short &amp; long-term disability insurance, life insurance, and various other discounts and perks.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_9b657c4e-8a1","directApply":true,"hiringOrganization":{"@type":"Organization","name":"xAI","sameAs":"https://www.x.ai/","logo":"https://logos.yubhub.co/x.ai.png"},"x-apply-url":"https://job-boards.greenhouse.io/xai/jobs/4803862007","x-work-arrangement":"hybrid","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$180,000 - $440,000 USD","x-skills-required":["distributed systems","stream processing","large-scale data platforms","Rust","Go","Scala","Kafka","Flink","Spark","Trino","Hadoop","debugging","profiling","performance optimization"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:40:03.394Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Palo Alto, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"distributed systems, stream processing, large-scale data platforms, Rust, Go, Scala, Kafka, Flink, Spark, Trino, Hadoop, debugging, profiling, performance optimization","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":180000,"maxValue":440000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_3367a9d1-967"},"title":"Engineering Manager , Data Engineering Solutions","description":"<p>We&#39;re looking for a manager to drive the Data Engineering Solutions Team in solving high-impact, cutting-edge data problems. The ideal candidate will be someone that has built data pipelines for large scale volume, is deeply knowledgeable of Data Engineering tools including Airflow/Spark/Kafka/Flink, is empathetic, excels at building strong relationships, and collaborates effectively with other Stripe teams to understand their use cases and unlock new capabilities.</p>\n<p>Key Responsibilities:</p>\n<ul>\n<li>Deliver cutting-edge data pipelines that scale to users&#39; needs, focusing on reliability and efficiency.</li>\n<li>Lead and manage a team of ambitious, talented engineers, providing mentorship, guidance, and support to ensure their success.</li>\n<li>Drive the execution of key reporting initiatives for Stripe, overseeing the entire development lifecycle from planning to delivery while maintaining high standards of quality and timely completion.</li>\n<li>Collaborate with product managers and key leaders across the company to create a shared roadmap and drive adoption of canonical datasets and data warehouses, use golden paths, and ensure Stripes are using trustworthy data.</li>\n<li>Understand user needs and pain points to prioritize engineering work and deliver high-quality solutions that meet user needs.</li>\n<li>Provide hands-on technical leadership in architecture/design, vision/direction/requirements setting, and incident response processes for your reports.</li>\n<li>Foster a collaborative and inclusive work environment, promoting innovation, knowledge sharing, and continuous improvement within the team.</li>\n<li>Partner with our recruiting team to attract and hire top talent, and define the overall hiring strategies for your team.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_3367a9d1-967","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Stripe","sameAs":"https://stripe.com/","logo":"https://logos.yubhub.co/stripe.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/stripe/jobs/7496118","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Airflow","Spark","Kafka","Flink","Data Engineering","Team Management","Leadership","Communication","Problem-Solving"],"x-skills-preferred":["Iceberg","Change Data Capture","Hive Metastore","Pinot","Trino","AWS Cloud"],"datePosted":"2026-03-31T18:12:23.063Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bengaluru"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Airflow, Spark, Kafka, Flink, Data Engineering, Team Management, Leadership, Communication, Problem-Solving, Iceberg, Change Data Capture, Hive Metastore, Pinot, Trino, AWS Cloud"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_b1d4c773-5c5"},"title":"Analytics Engineer, Finance","description":"<p><strong>Compensation</strong></p>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>The Finance Data team is embedded within the CFO Org and is responsible for building internal data products that scale analytics across business teams and drive efficiencies in our daily operations. This team provides technical guidance on high-impact, scalable projects across Finance, and is the subject-matter expert in financial and transactional data that supports our Finance day-to-day operations.</p>\n<p><strong>About the Role</strong></p>\n<p>As an Analytics Engineer, you will be setting the foundation to scale analytics across our business functions and impart best data practices for a rapidly growing organization. We aspire to build the Finance team of the future.</p>\n<p>In addition, you will work collaboratively with key stakeholders in Finance and other business teams to understand their pain points and take the lead in proposing viable, future-proof solutions to resolve them. You will also autonomously lead your own projects that deliver business impact and help cultivate a mature data culture among Finance teams.</p>\n<p>We are looking for a seasoned engineer who has a proven track record of owning the entire data stack at high transaction volume companies, managing business critical ETL pipelines consumed by non-technical teams. As a generalist “fixer”, you may be deployed across several different Finance domains (e.g. Tax datamart, ERP migration, Procurement automation). For this role we need someone who excels in dynamic environments, adapts quickly to changing needs, and confidently navigates ambiguous or evolving requirements. If you&#39;re energized by solving technical problems without a playbook and comfortable wearing multiple hats, this role is for you! To clarify, you will <strong>not</strong> be responsible for training ML models and neither would we describe this role as ‘product analytics’.</p>\n<p>This role is based in San Francisco, CA. We use a hybrid work model of 3 days in the office per week and offer relocation assistance to new employees.</p>\n<p><strong>In this role, you will:</strong></p>\n<ul>\n<li>Understand the data needs of Finance teams, including Revenue, Tax, Procurement, Compute &amp; Infrastructure Accounting, Strategic Finance, and translate that scope into technical requirements</li>\n</ul>\n<ul>\n<li>Facilitate the development of data products and tools to for stakeholders to self-service and enable analytics to scale across the company</li>\n</ul>\n<ul>\n<li>Lead dimensional design - define, own, and maintain business facing data marts</li>\n</ul>\n<ul>\n<li>Be a cross-functional champion at upholding high data integrity standards and SLAs for the timely delivery of data</li>\n</ul>\n<ul>\n<li>Build and maintain insightful and reliable dashboards to track both operational and financial Metrics for the Executive team</li>\n</ul>\n<ul>\n<li>Contribute to the future roadmap of the Finance team from a data systems perspective</li>\n</ul>\n<ul>\n<li>Grow to be an expert in Finance Data and OpenAI’s data architecture</li>\n</ul>\n<p><strong>You might thrive in this role if you have:</strong></p>\n<ul>\n<li>7+ years of experience as an Analytics Engineer or in a similar role (Data Analyst or Data Engineer) with a proven track record in shipping canonical datasets</li>\n</ul>\n<ul>\n<li>Empathy towards non-developer stakeholders and their day-to-day pain points</li>\n</ul>\n<ul>\n<li>Strong proficiency in SQL for data transformation, comfort in at least one functional/OOP language such as Python or R</li>\n</ul>\n<ul>\n<li>Familiarity with managing distributed data stores (e.g. S3, Trino, Hive, Spark), and experience building multi-step ETL jobs coupled with orchestrating workflows (e.g. Airflow, Dagster)</li>\n</ul>\n<ul>\n<li>Experience in writing unit tests to validate data products and version control (e.g. GitHub, Stash)</li>\n</ul>\n<ul>\n<li>Expert at creating compelling data visualizations with dashboarding tools (e.g. Tableau, Looker or similar)</li>\n</ul>\n<ul>\n<li>Excellent communication skills and ability to present data-driven narratives in both verbal and written form to a non-technical audience</li>\n</ul>\n<ul>\n<li>Experience solving ambiguous problem statements in an early stage environment</li>\n</ul>\n<p><strong>You could be an especially great fit if you have:</strong></p>\n<ul>\n<li>Prior experience leading the development of an internal production tool, serving hundreds of cross-functional customers such as Billing Operations, Deal Desk or Go-to-Market teams</li>\n</ul>\n<ul>\n<li>Some frontend experience with React, TypeScript, Retool, Streamlit, or building web apps</li>\n</ul>\n<ul>\n<li>Good understanding of Spark and ability to write, debug, and optimize Spark jobs</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>\n<p>We are an equal opportunity employer, and we do not discriminate on the basis of race, religion, color, national origin, sex, sexual orientation, age, veteran status, disability, genetic information, or other applicable legally protected characteristic.</p>\n<p>For additional information, please see [OpenAI’s Affirmative Action and Equal Employment Opportunity Policy Statement](https://cdn.openai.com/policies/eeo-policy-statement.pdf).</p>\n<p>Background checks for applicants will be administered in accordance with applicable law, and qualified applicants with arrest or conviction records will be considered for employment consistent with those laws, including the San Francisco Fair Chance Ordinance, the Los Angeles County Fair Chance Ordinance for Employers, and the California Fair Chance Act, for US-based candidates. For unincorporated Los Angeles County workers: we reasonably believe that criminal history may have a direct, adverse and negative relationship with the following job duties, potentially resulting in the withdrawal of a conditional offer of employment: protect computer hardware entrusted to you from theft, loss or damage; return all computer</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_b1d4c773-5c5","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/7cd50a19-65f2-4a52-89a2-512130e58c5c","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"Full time","x-salary-range":"$198K – $260K • Offers Equity","x-skills-required":["SQL","Python","R","S3","Trino","Hive","Spark","Airflow","Dagster","GitHub","Stash","Tableau","Looker"],"x-skills-preferred":["React","TypeScript","ReTool","Streamlit","Web development"],"datePosted":"2026-03-08T22:16:37.388Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, R, S3, Trino, Hive, Spark, Airflow, Dagster, GitHub, Stash, Tableau, Looker, React, TypeScript, ReTool, Streamlit, Web development","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":198000,"maxValue":260000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8c9ae282-129"},"title":"Member of Technical Staff - Data Platform","description":"<p><strong>Summary</strong></p>\n<p>Microsoft are looking for a talented Member of Technical Staff - Data Platform at their Mountain View office. This role sits at the heart of designing distributed systems that process petabytes of data for the world&#39;s most advanced AI models. You will own the platform that transforms raw, massive-scale signals into the fuel that powers training, inference, and evaluation for millions of users.</p>\n<p><strong>About the Role</strong></p>\n<p>As a Member of Technical Staff - Data Platform, you will be responsible for designing and building the underlying frameworks that allow internal teams to process massive datasets efficiently, abstracting away the complexity of &#39;ETL&#39; into self-service infrastructure. You will modernize our data stack by moving from batch-heavy patterns to event-driven architectures, utilizing modern streaming architecture to reduce latency for AI inference.</p>\n<p><strong>Accountabilities</strong></p>\n<ul>\n<li>Design and build the underlying frameworks that allow internal teams to process massive datasets efficiently, abstracting away the complexity of &#39;ETL&#39; into self-service infrastructure.</li>\n<li>Modernize our data stack by moving from batch-heavy patterns to event-driven architectures, utilizing modern streaming architecture to reduce latency for AI inference.</li>\n</ul>\n<p><strong>The Candidate we&#39;re looking for</strong></p>\n<p><strong>Experience:</strong></p>\n<ul>\n<li>Master&#39;s Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 3+ years experience in business analytics, data science, software development, data modeling, or data engineering OR Bachelor&#39;s Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 4+ years experience in business analytics, data science, software development, data modeling, or data engineering OR equivalent experience.</li>\n</ul>\n<p><strong>Technical skills:</strong></p>\n<ul>\n<li>Proficiency in Python, Scala, Java, or Go.</li>\n<li>Deep Distributed Systems Knowledge: Demonstrated technical understanding of massive-scale compute engines (e.g., Apache Spark, Flink, Ray, Trino, or Snowflake).</li>\n</ul>\n<p><strong>Personal attributes:</strong></p>\n<ul>\n<li>Strong background in streaming technologies (Kafka, Azure EventHubs, Pulsar) and stateful stream processing.</li>\n<li>Experience with container orchestration (Kubernetes) for deploying data applications.</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Competitive salary range: $119,800 - $234,700 per year.</li>\n<li>Comprehensive benefits package, including medical, dental, and vision insurance.</li>\n<li>401(k) matching program.</li>\n<li>Paid time off and holidays.</li>\n<li>Opportunities for professional growth and development.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8c9ae282-129","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Microsoft","sameAs":"https://microsoft.ai","logo":"https://logos.yubhub.co/microsoft.ai.png"},"x-apply-url":"https://microsoft.ai/job/member-of-technical-staff-data-platform/","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"$119,800 - $234,700 per year","x-skills-required":["Python","Scala","Java","Go","Apache Spark","Flink","Ray","Trino","Snowflake"],"x-skills-preferred":["Kafka","Azure EventHubs","Pulsar","Kubernetes"],"datePosted":"2026-03-06T07:26:42.738Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Mountain View"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Scala, Java, Go, Apache Spark, Flink, Ray, Trino, Snowflake, Kafka, Azure EventHubs, Pulsar, Kubernetes","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":119800,"maxValue":234700,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4c6c34df-ee0"},"title":"Member of Technical Staff - Data Platform","description":"<p><strong>Summary</strong></p>\n<p>Microsoft are looking for a talented Member of Technical Staff - Data Platform at their Redmond office. This role sits at the heart of designing distributed systems that process petabytes of data for the world&#39;s most advanced AI models. You will own the platform that transforms raw, massive-scale signals into the fuel that powers training, inference, and evaluation for millions of users.</p>\n<p><strong>About the Role</strong></p>\n<p>As a Member of Technical Staff - Data Platform, you will be responsible for designing and building the underlying frameworks that allow internal teams to process massive datasets efficiently, abstracting away the complexity of &#39;ETL&#39; into self-service infrastructure. You will modernize our data stack by moving from batch-heavy patterns to event-driven architectures, utilizing modern streaming architecture to reduce latency for AI inference.</p>\n<p><strong>Accountabilities</strong></p>\n<ul>\n<li>Design and build the underlying frameworks that allow internal teams to process massive datasets efficiently</li>\n<li>Modernize our data stack by moving from batch-heavy patterns to event-driven architectures</li>\n</ul>\n<p><strong>The Candidate we&#39;re looking for</strong></p>\n<p><strong>Experience:</strong></p>\n<ul>\n<li>Master&#39;s Degree in Computer Science, Math, Software Engineering, Computer Engineering, or related field AND 3+ years experience in business analytics, data science, software development, data modeling, or data engineering</li>\n</ul>\n<p><strong>Technical skills:</strong></p>\n<ul>\n<li>Proficiency in Python, Scala, Java, or Go</li>\n<li>Deep Distributed Systems Knowledge: Demonstrated technical understanding of massive-scale compute engines (e.g., Apache Spark, Flink, Ray, Trino, or Snowflake)</li>\n</ul>\n<p><strong>Personal attributes:</strong></p>\n<ul>\n<li>Strong background in streaming technologies (Kafka, Azure EventHubs, Pulsar) and stateful stream processing</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Competitive salary</li>\n<li>Comprehensive benefits package</li>\n<li>Opportunities for professional growth and development</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4c6c34df-ee0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Microsoft","sameAs":"https://microsoft.ai","logo":"https://logos.yubhub.co/microsoft.ai.png"},"x-apply-url":"https://microsoft.ai/job/member-of-technical-staff-data-platform-2/","x-work-arrangement":"onsite","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":"USD $119,800 – $234,700 per year","x-skills-required":["Python","Scala","Java","Go","Apache Spark","Flink","Ray","Trino","Snowflake"],"x-skills-preferred":["Kafka","Azure EventHubs","Pulsar"],"datePosted":"2026-03-06T07:25:59.144Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Redmond"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Scala, Java, Go, Apache Spark, Flink, Ray, Trino, Snowflake, Kafka, Azure EventHubs, Pulsar","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":119800,"maxValue":234700,"unitText":"YEAR"}}}]}