{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/title/senior-data-engineer"},"x-facet":{"type":"title","slug":"senior-data-engineer","display":"Senior Data Engineer","count":9},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e58b08f7-c31"},"title":"Senior Data Engineer","description":"<p>As a Senior Data Engineer on the Analytics Team, you will collaborate with stakeholders across the company to design, build and implement data pipelines and models that enable our next generation of technology to be deployed around the world. You will have a hand in helping shape the data platform vision at Anduril.</p>\n<p>We&#39;re looking for software and data engineers who are seeking high impact collaborative roles focused on driving operational execution. Ideally you are looking to learn what it takes to build the next generation of defence technology.</p>\n<p>Your responsibilities will include leading the design and roadmap for our data platform, partnering with operations, product, and engineering to advocate best practices and build supporting systems and infrastructure for the various data needs, owning the ingest and egress frameworks for data pipelines that stitch together various data sources in order to produce valuable data products that drive the business, and managing a large user base and providing true data self-service at scale.</p>\n<p>We use Palantir Foundry as our central hub for data-driven applications, visualizations and large-scale data analysis across the Anduril org. We also use SQLMesh for data transformations, Athena for querying data, Apache Iceberg as our table format, and Flyte for orchestration.</p>\n<p>Required qualifications include 5+ years of experience in a data engineering role building products, ideally in a fast-paced environment, good foundations in Python or another language, experience with Spark, PySpark, SQL and dbt, experience with Enterprise Data Systems like Palantir Foundry, and experience with or interest in learning how to develop data services and data products.</p>\n<p>The salary range for this role is $166,000-$220,000 USD.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e58b08f7-c31","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anduril","sameAs":"https://www.anduril.com/","logo":"https://logos.yubhub.co/anduril.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/andurilindustries/jobs/4587312007","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$166,000-$220,000 USD","x-skills-required":["Python","Spark","PySpark","SQL","dbt","Palantir Foundry","SQLMesh","Athena","Apache Iceberg","Flyte"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:58:44.003Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Costa Mesa, California, United States"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Spark, PySpark, SQL, dbt, Palantir Foundry, SQLMesh, Athena, Apache Iceberg, Flyte","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":166000,"maxValue":220000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_a98d4ace-d27"},"title":"Senior Data Engineer","description":"<p>We are looking for a Senior Data Engineer to join our high-performing data enablement team. As a Senior Data Engineer, you will play a pivotal role within the Data team that powers Yuno and its payment platform, while helping co-design and implement an architecture that scales with the product and the company.</p>\n<p>The stack is modern: StarRocks as our primary analytical layer, Flink for processing, DBT for transformation, Airflow for orchestration and various tooling for surfacing insights.</p>\n<p>You&#39;ll be working on things that matter and are technically interesting:</p>\n<ul>\n<li><p>Design and build data pipelines for large volumes of payment data that are performant, reliable, and correct , not just fast.</p>\n</li>\n<li><p>Own end-to-end data flows: from ingestion and transformation through to the outputs that Finance, Product, and clients depend on.</p>\n</li>\n<li><p>Drive data quality across your domain with tooling.</p>\n</li>\n<li><p>Work cross-functionally with Product, Finance and enable other Engineering teams via a &#39;consulting&#39; style model.</p>\n</li>\n<li><p>Contribute to how the team works , code review culture, CI/CD standards, ADRs, how we handle incidents , we&#39;re building these practices now and senior engineers shape them.</p>\n</li>\n<li><p>Help onboard and level up engineers around you; there&#39;s real opportunity to make an impact here.</p>\n</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_a98d4ace-d27","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Yuno","sameAs":"https://www.yuno.com/","logo":"https://logos.yubhub.co/yuno.com.png"},"x-apply-url":"https://jobs.lever.co/yuno/dc30ae7b-9c0f-426f-ae77-c58d9e4f6d6d","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Proven proactivity, technical acumen and the ability to lead initiatives and deliver projects.","Experience in defining and evolving data engineering standards, architectural guidelines and governance, ideally within a regulated environment.","Strong Python and SQL skills.","Hands-on experience with Spark or Flink in production.","DBT for data transformation."],"x-skills-preferred":["Airflow for orchestration.","Experience with Apache Hudi.","Experience with financial, transactional, or payment data."],"datePosted":"2026-04-17T13:11:45.302Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Europe"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Proven proactivity, technical acumen and the ability to lead initiatives and deliver projects., Experience in defining and evolving data engineering standards, architectural guidelines and governance, ideally within a regulated environment., Strong Python and SQL skills., Hands-on experience with Spark or Flink in production., DBT for data transformation., Airflow for orchestration., Experience with Apache Hudi., Experience with financial, transactional, or payment data."},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_11a36eab-3cb"},"title":"Senior Data Engineer","description":"<p><strong>Job Description</strong></p>\n<p>Are you ready to contribute to the evolution of our data pipelines for our B2C division? At Future, we are transforming our data-driven decision-making processes and we are looking for a passionate and experienced Data Engineer to join us.</p>\n<p>This is an exciting opportunity for someone who excels in a creative environment, enjoys solving complex data challenges, and is eager to build impactful business insights, for this role you will directly report into the Head of Data Engineering</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Develop and maintain new/current features of the data platform.</li>\n<li>Responsible for delivery of development projects, including scoping, writing and sizing of stories involved.</li>\n<li>Take ownership of BAU processes, develop area specific domain mastery, and seek means to automate them or reduce their impact.</li>\n<li>Proposes and advocates for changes to reduce risk, cost and overhead.</li>\n<li>Provide appropriate documentation for pipelines developed</li>\n<li>Parameterise pipelines so configuration can be changed easily without having to perform deep changes to the codebase</li>\n<li>Apply appropriate testing principles to ensure code is fit for purpose</li>\n</ul>\n<p><strong>Experience</strong></p>\n<ul>\n<li>Experience using Python on Google Cloud Platform for Big Data projects, BigQuery, DataFlow (Apache Beam), Cloud Run Functions, Cloud Run, Cloud Workflows, Cloud Composure</li>\n<li>SQL development skills</li>\n<li>Experience using Dataform or dbt</li>\n<li>Demonstrated strength in data modelling, ETL development, and data warehousing</li>\n<li>Knowledge of data management fundamentals and data storage principles</li>\n<li>Familiarity with statistical models or data mining algorithms and practical experience applying these to business problems</li>\n</ul>\n<p><strong>What&#39;s in it for you</strong></p>\n<p>The expected range for this role is £50,000 - £60,000</p>\n<p>This is a Hybrid role from our Bath Office, working three days from the office, two from home … Plus more great perks, which include;</p>\n<ul>\n<li>Uncapped leave, because we trust you to manage your workload and time</li>\n<li>When we hit our targets, enjoy a share of our profits with a bonus</li>\n<li>Refer a friend and get rewarded when they join Future</li>\n<li>Wellbeing support with access to our Colleague Assistant Programmes</li>\n<li>Opportunity to purchase shares in Future, with our Share Incentive Plan</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_11a36eab-3cb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Future","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/3535C2B9B5","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"£50,000 - £60,000","x-skills-required":["Python","Google Cloud Platform","BigQuery","DataFlow","Apache Beam","Cloud Run Functions","Cloud Run","Cloud Workflows","Cloud Composure","SQL","Dataform","dbt","data modelling","ETL development","data warehousing","data management fundamentals","data storage principles","statistical models","data mining algorithms"],"x-skills-preferred":[],"datePosted":"2026-03-09T16:21:59.655Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Bath"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Google Cloud Platform, BigQuery, DataFlow, Apache Beam, Cloud Run Functions, Cloud Run, Cloud Workflows, Cloud Composure, SQL, Dataform, dbt, data modelling, ETL development, data warehousing, data management fundamentals, data storage principles, statistical models, data mining algorithms","baseSalary":{"@type":"MonetaryAmount","currency":"GBP","value":{"@type":"QuantitativeValue","minValue":50000,"maxValue":60000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_5d911052-764"},"title":"Senior Data Engineer","description":"<p><strong>About the Role</strong></p>\n<p>We&#39;re hiring a Senior Data Engineer to work on our Data Lake Team. As a key member of the team, you will be responsible for building and operating various data platform components, including data quality, data pipelines, infrastructure, and monitoring.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Maintain data pipeline job framework</li>\n<li>Develop Data Quality framework ( internal set of tools for internal and external data sources validation )</li>\n<li>Maintain and develop public facing data ingestion service with 17 000+ RPS.</li>\n<li>Maintain and develop core data pipelines in batch and streaming manners.</li>\n<li>Be a last line of support for our internal platform users.</li>\n<li>Take a part in an on-call rotation for data platform incidents (shared across the team).</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<ul>\n<li>Fluent English</li>\n<li>4+ years building production services and data pipelines (batch and/or streaming)</li>\n<li>Strong experience with Python or the readiness to ramp up quickly.</li>\n<li>Hands-on experience with at least one MPP system (Spark, Trino, Redshift etc.)</li>\n<li>Hands-on experience operating services in a cloud environment (AWS preferred)</li>\n</ul>\n<p><strong>Nice to Have</strong></p>\n<ul>\n<li>Terraform/CloudFormation or other IaC tools</li>\n<li>ClickHouse or similar analytical databases</li>\n<li>Experiences with data quality/observability tools</li>\n</ul>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Unlimited vacation time - we strongly encourage all employees to take at least 3 weeks per year</li>\n<li>Fully remote team - choose where you live</li>\n<li>Work from home stipend - we want you to have the resources you need to set up your home office</li>\n<li>Apple laptops provided for new employees</li>\n<li>Training and development budget - refreshed each year for every employee</li>\n<li>Maternity &amp; Paternity leave for qualified employees</li>\n<li>Work with smart people who will help you grow and make a meaningful impact</li>\n<li>Base salary: $80k–$120k USD, depending on knowledge, skills, experience, and interview results</li>\n<li>Stock options - offered in addition to the base salary</li>\n<li>Regular team offsites to connect and collaborate</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_5d911052-764","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/FF201D8AA3","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$80k–$120k USD","x-skills-required":["Python","MPP system","AWS"],"x-skills-preferred":["Terraform","ClickHouse","data quality/observability tools"],"datePosted":"2026-03-09T10:57:58.178Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, MPP system, AWS, Terraform, ClickHouse, data quality/observability tools","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":120000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_70fe3dd2-f85"},"title":"Senior Data Engineer","description":"<p><strong>About the Role</strong></p>\n<p>We&#39;re hiring a Senior Data Engineer to work on our Data Infrastructure Team. This team is responsible for building and maintaining the Data Platform, a comprehensive set of tools and infrastructure used daily by every data scientist and ML engineer in our company.</p>\n<p><strong>Responsibilities</strong></p>\n<ul>\n<li>Job scheduling and orchestration for data pipelines.</li>\n<li>Deployment and management of BI tools.</li>\n<li>Real-time analytics infrastructure (ClickHouse, AWS Lambda, Cube.js, and related tooling).</li>\n<li>Real-time log ingestion and processing, including data compliance.</li>\n<li>Core data services (e.g., Kubernetes, Ray, metadata services) and enterprise-wide observability solutions (based on ClickHouse and OpenTelemetry).</li>\n</ul>\n<p><strong>Requirements</strong></p>\n<p>We are seeking an engineer with at least 4 years of experience who possesses strong programming skills (ideally in Python), and expertise in big data engineering, web services, and cloud platforms (ideally AWS). We are looking for someone eager to build diverse components and drive the evolution of our platform while working closely with our users. Excellent English communication skills and robust computer science background is a strong requirement.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Unlimited vacation time - we strongly encourage all of our employees take at least 3 weeks per year</li>\n<li>Fully remote team - choose where you live</li>\n<li>Work from home stipend! We want you to have the resources you need to set up your home office</li>\n<li>Apple laptops provided for new employees</li>\n<li>Training and development budget for every employee, refreshed each year</li>\n<li>Maternity &amp; Paternity leave for qualified employees</li>\n<li>Work with smart people who will help you grow and make a meaningful impact</li>\n<li>This position has a base salary range between $80k and $120k USD. The offer varies on many factors including job related knowledge, skills, experience, and interview results.</li>\n<li>Regular team offsites to connect and collaborate</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_70fe3dd2-f85","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Constructor","sameAs":"https://apply.workable.com","logo":"https://logos.yubhub.co/j.com.png"},"x-apply-url":"https://apply.workable.com/j/C6407C4CB5","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$80k - $120k USD","x-skills-required":["Python","big data engineering","web services","cloud platforms (AWS)"],"x-skills-preferred":["ClickHouse","AWS Lambda","Cube.js"],"datePosted":"2026-03-09T10:57:40.511Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, big data engineering, web services, cloud platforms (AWS), ClickHouse, AWS Lambda, Cube.js","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":80000,"maxValue":120000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_015e5c6d-a31"},"title":"Senior Data Engineer","description":"<p><strong>Why Valvoline Global Operations?</strong></p>\n<p>At Valvoline Global Operations, we&#39;re proud to be The Original Motor Oil, but we&#39;ve never rested on being first. Founded in 1866, we introduced the world&#39;s first branded motor oil, staking our claim as a pioneer in the automotive and industrial solutions industry.</p>\n<p><strong>Job Purpose</strong></p>\n<p>We are seeking a highly skilled and motivated Data Engineer to join our growing data and analytics team. The ideal candidate will have strong experience designing and developing scalable data pipelines, integrating complex systems, and optimizing data workflows. Proficiency in Databricks and SAP Datasphere is preferred, as these platforms are central to our data ecosystem.</p>\n<p><strong>How You Make an Impact (Job Accountabilities)</strong></p>\n<ul>\n<li>Design, build, and maintain robust, scalable, and high-performance data pipelines using Databricks and SAP Datasphere.</li>\n<li>Collaborate with data architects, analysts, data scientists, and business stakeholders to gather requirements and deliver data solutions aligned with stakeholders&#39; goals.</li>\n<li>Integrate diverse data sources (e.g., SAP, APIs, flat files, cloud storage) into the enterprise data platforms</li>\n<li>Ensure high standards of data quality and implement data governance practices. Stay current with emerging trends and technologies in cloud computing, big data, and data engineering.</li>\n<li>Provide ongoing support for the platform, troubleshoot any issues that arise, and ensure high availability and reliability of data infrastructure.</li>\n<li>Create documentation for the platform infrastructure and processes, and train other team members or users in platform effectively.</li>\n</ul>\n<p><strong>What You Bring to the Role (Job Qualifications / Education / Skills / Requirements / Capabilities)</strong></p>\n<ul>\n<li>Bachelor&#39;s or Master’s degree in Computer Science, Data Engineering, Information Systems, or a related field.</li>\n<li>5-7+ years of experience in a data engineering or related role.</li>\n<li>Strong knowledge of data engineering principles, data warehousing concepts, and modern data architecture.</li>\n<li>Proficiency in SQL and at least one programming language (e.g., Python, Scala).</li>\n<li>Experience with cloud platforms (e.g., Azure, AWS, or GCP), particularly in data services.</li>\n<li>Familiarity with data orchestration tools (e.g., PySpark, Airflow, Azure Data Factory) and CI/CD pipelines.</li>\n</ul>\n<p><strong>Competencies Desired</strong></p>\n<ul>\n<li>Hands-on experience with Databricks (including Spark/PySpark, Delta Lake, MLflow, Unity Catalog, etc.).</li>\n<li>Practical experience working with SAP Datasphere (or SAP Data Warehouse Cloud) in data modeling and data integration scenarios.</li>\n<li>SAP BW or SAP HANA experience is a plus.</li>\n<li>Experience with BI tools like Power BI or Tableau.</li>\n<li>Understanding of data governance frameworks and data security best practices.</li>\n<li>Exposure to data lakehouse architecture and real-time streaming data pipelines.</li>\n<li>Certifications in Databricks, SAP, or cloud platforms are advantageous.</li>\n</ul>\n<p><strong>Working Conditions / Physical Requirements / Travel Requirements</strong></p>\n<ul>\n<li>Normal Office environment.</li>\n<li>Prolonged periods of computer use and frequent participation in meetings</li>\n<li>Occasional walking, standing, and light lifting (up to 10 lbs)</li>\n</ul>\n<ul>\n<li>Minimal travel required.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_015e5c6d-a31","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Valvoline Global Operations","sameAs":"https://jobs.valvolineglobal.com","logo":"https://logos.yubhub.co/jobs.valvolineglobal.com.png"},"x-apply-url":"https://jobs.valvolineglobal.com/job/Senior-Data-Engineer/1316654400/","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["data engineering","Databricks","SAP Datasphere","SQL","Python","Scala","cloud platforms","data orchestration tools","CI/CD pipelines"],"x-skills-preferred":["Databricks","SAP Datasphere","SAP BW","SAP HANA","Power BI","Tableau","data governance frameworks","data security best practices","data lakehouse architecture","real-time streaming data pipelines"],"datePosted":"2026-03-08T22:14:37.507Z","jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Automotive","skills":"data engineering, Databricks, SAP Datasphere, SQL, Python, Scala, cloud platforms, data orchestration tools, CI/CD pipelines, Databricks, SAP Datasphere, SAP BW, SAP HANA, Power BI, Tableau, data governance frameworks, data security best practices, data lakehouse architecture, real-time streaming data pipelines"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_63e3e256-1a7"},"title":"Senior Data Engineer","description":"<p><strong>Senior Data Engineer</strong></p>\n<p><strong>Location</strong></p>\n<p>London</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Location Type</strong></p>\n<p>Hybrid</p>\n<p><strong>Department</strong></p>\n<p>CommercialRevenue Operations</p>\n<p>Synthesia is the world&#39;s leading AI video platform for business, used by over 90% of the Fortune 100. Founded in 2017, the company is headquartered in London, with offices and teams across Europe and the US.</p>\n<p>As AI continues to shape the way we live and work, Synthesia develops products to enhance visual communication and enterprise skill development, helping people work better and stay at the centre of successful organisations.</p>\n<p>Following our recent Series E funding round, where we raised $200 million, our valuation stands at $4 billion. Our total funding exceeds $530 million from premier investors including Accel, NVentures (Nvidia&#39;s VC arm), Kleiner Perkins, GV, and Evantic Capital, alongside the founders and operators of Stripe, Datadog, Miro, and Webflow.</p>\n<p><strong>Senior Data Engineer</strong></p>\n<p>We&#39;re hiring a Senior Data Engineer to join Synthesia and take ownership of our core data systems. You&#39;ll be responsible for designing and maintaining scalable pipelines, optimising data models, and ensuring high data quality and governance standards.</p>\n<p><strong>What you&#39;ll do at Synthesia:</strong></p>\n<ul>\n<li>Architect and scale robust, end-to-end data pipelines that ingest and transform complex semi-structured and structured data into our Snowflake data warehouse.</li>\n</ul>\n<ul>\n<li>Own the evolution of our dbt project - implementing modular modelling patterns and other best practices to ensure a &#39;single source of truth&#39; for the entire organisation.</li>\n</ul>\n<ul>\n<li>Manage platform infrastructure in snowflake, AWS and other tools.</li>\n</ul>\n<ul>\n<li>Continuously optimise warehouse performance and cost by diagnosing bottlenecks, tuning inefficient queries, and improving how compute resources are used as we scale.</li>\n</ul>\n<ul>\n<li>Bridge the gap between experimental data science workflows and production, building the infrastructure and orchestration needed to deploy and monitor batch ML jobs.</li>\n</ul>\n<ul>\n<li>Drive best practices in data security, governance, and compliance, particularly with regards to AI.</li>\n</ul>\n<ul>\n<li>Partner with cross-functional stakeholders to understand data requirements and translate them into technical solutions.</li>\n</ul>\n<p><strong>What we&#39;re looking for:</strong></p>\n<ul>\n<li>5+ years of experience as a Data Engineer or in a closely related role, with a proven track record of building and operating production data systems.</li>\n</ul>\n<ul>\n<li>Experience working in an early-stage or scaling data function. You&#39;re comfortable taking ownership and wearing multiple hats when needed.</li>\n</ul>\n<ul>\n<li>Strong foundations in software engineering and data modelling best practices, with an ability to design systems that are maintainable, scalable, and easy for others to build on.</li>\n</ul>\n<ul>\n<li>Deep expertise in SQL, and solid experience using Python or similar languages to build data pipelines, tooling, and orchestration (Airflow).</li>\n</ul>\n<ul>\n<li>Hands on experience managing cloud infrastructure using infrastructure-as-code (e.g. Terraform) on AWS, GCP, or similar platforms.</li>\n</ul>\n<ul>\n<li>A pragmatic approach to data platform design, with an eye for performance, cost efficiency, and operational reliability.</li>\n</ul>\n<ul>\n<li>Excellent communication skills: you can work effectively with technical and non-technical stakeholders to gather requirements, explain trade-offs and communicate data team needs.</li>\n</ul>\n<ul>\n<li>A product-oriented mindset, with an understanding of how data can shape decision making and accelerate company growth.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_63e3e256-1a7","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Synthesia","sameAs":"https://www.synthesia.io/","logo":"https://logos.yubhub.co/synthesia.io.png"},"x-apply-url":"https://jobs.ashbyhq.com/synthesia/46650970-494a-4d4b-ab4b-75c2a3b06daf","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["SQL","Python","Airflow","Terraform","AWS","GCP","Snowflake","dbt"],"x-skills-preferred":[],"datePosted":"2026-03-06T18:32:20.724Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"London"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"SQL, Python, Airflow, Terraform, AWS, GCP, Snowflake, dbt"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4a7597fd-d7a"},"title":"Senior Data Engineer","description":"<p>Joining Razer will place you on a global mission to revolutionize the way the world games. Razer is a place to do great work, offering you the opportunity to make an impact globally while working across a global team located across 5 continents. Razer is also a great place to work, providing you the unique, gamer-centric #LifeAtRazer experience that will put you in an accelerated growth, both personally and professionally.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference. You will collaborate closely with AI/ML engineers, data scientists, and platform teams to build the next generation of AI-driven products.</p>\n<ul>\n<li>Lead AI Data Engineering initiatives by driving the design and development of robust data pipelines for AI/ML workloads, ensuring efficiency, scalability, and reliability.</li>\n<li>Design and implement data architectures that support AI model training, including feature stores, vector databases, and real-time streaming solutions.</li>\n<li>Develop high performance data pipelines that process structured, semi-structured, and unstructured data at scale, supporting the various AI applications</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Hands on experience working with Vector/Graph;Neo4j</li>\n<li>3+ years of experience in data engineering, working on AI/ML-driven data architectures</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4a7597fd-d7a","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Razer","sameAs":"https://razer.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/razer.com.png"},"x-apply-url":"https://razer.wd3.myworkdayjobs.com/en-US/Careers/job/Singapore/Senior-Data-Engineer_JR2025005485","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Hands on experience working with Vector/Graph;Neo4j","3+ years of experience in data engineering, working on AI/ML-driven data architectures"],"x-skills-preferred":["Python","SQL","Experience in developing and deploying applications running on cloud infrastructure such as AWS, Azure or Google Cloud Platform using Infrastructure as code tools such as Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes","Experience using orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT)","Excellent with various data processing techniques (both streaming and batch), managing and optimizing data storage (Data Lake, Lake House and Database, SQL, and NoSQL) is essential."],"datePosted":"2026-01-01T15:49:59.491Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Hands on experience working with Vector/Graph;Neo4j, 3+ years of experience in data engineering, working on AI/ML-driven data architectures, Python, SQL, Experience in developing and deploying applications running on cloud infrastructure such as AWS, Azure or Google Cloud Platform using Infrastructure as code tools such as Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes, Experience using orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT), Excellent with various data processing techniques (both streaming and batch), managing and optimizing data storage (Data Lake, Lake House and Database, SQL, and NoSQL) is essential."},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_e5eb908e-6f9"},"title":"Senior Data Engineer","description":"<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference.</p>\n<p><strong>What you&#39;ll do</strong></p>\n<p>We are looking for a Senior Data Engineer to lead the technical initiatives for AI Data Engineering, enabling scalable, high-performance data pipelines that power AI and machine learning applications. This role will focus on architecting, optimizing, and managing data infrastructure to support AI model training, feature engineering, and real-time inference.</p>\n<ul>\n<li>Lead AI Data Engineering initiatives by driving the design and development of robust data pipelines for AI/ML workloads, ensuring efficiency, scalability, and reliability.</li>\n<li>Design and implement data architectures that support AI model training, including feature stores, vector databases, and real-time streaming solutions.</li>\n</ul>\n<p><strong>What you need</strong></p>\n<ul>\n<li>Hands on experience working with Vector/Graph;Neo4j</li>\n<li>3+ years of experience in data engineering, working on AI/ML-driven data architectures</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_e5eb908e-6f9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Razer","sameAs":"https://razer.wd3.myworkdayjobs.com","logo":"https://logos.yubhub.co/razer.com.png"},"x-apply-url":"https://razer.wd3.myworkdayjobs.com/en-US/Careers/job/Singapore/Senior-Data-Engineer_JR2025005485","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Vector/Graph;Neo4j","data engineering","AI/ML-driven data architectures"],"x-skills-preferred":["Python","SQL","Terraform","containerization tools like Dockers","container orchestration platforms like Kubernetes","orchestration tools like Airflow or Prefect","distributed computing framework like Spark or Dask","data transformation tool like Data Build Tool (DBT)"],"datePosted":"2025-12-26T10:53:07.867Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Singapore"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Vector/Graph;Neo4j, data engineering, AI/ML-driven data architectures, Python, SQL, Terraform, containerization tools like Dockers, container orchestration platforms like Kubernetes, orchestration tools like Airflow or Prefect, distributed computing framework like Spark or Dask, data transformation tool like Data Build Tool (DBT)"}]}