<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>8f03ad2d-96f</externalid>
      <Title>Software Engineer, Research Data Platform</Title>
      <Description><![CDATA[<p>We&#39;re looking for engineers who love working directly with users and who excel at building data products. The Research Data Platform team builds the tools that Anthropic&#39;s researchers use every day to manage, query, and analyze the data that goes into training and evaluating frontier models.</p>
<p>As a Software Engineer on the Research Data Platform team, you will:</p>
<ul>
<li>Build and operate data pipelines that extract data from research training runs and land it in storage systems that are easy and fast to query</li>
<li>Work closely with researchers to design and build APIs, libraries, and web interfaces that support data management, exploration, and analysis</li>
<li>Develop dataset management, data cataloging, and provenance tooling that researchers use in their day-to-day work</li>
<li>Embed with research teams to understand their workflows, identify high-leverage tooling opportunities, and ship solutions quickly</li>
<li>Collaborate with adjacent teams to build on existing systems rather than reinventing them</li>
</ul>
<p>We do not require prior ML or AI training experience. If you enjoy working closely with technical users, learning new domains quickly, and building tools people actually want to use, you&#39;ll pick up the research context fast.</p>
<p>Strong candidates may also have experience with large-scale ETL, columnar storage formats, and query engines (e.g., Spark, BigQuery, DuckDB, Parquet), high-volume time series data , ingestion, storage, and efficient querying, data cataloging, lineage, or metadata management systems, or ML experiment tracking or metrics platforms.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$320,000-$405,000 USD</Salaryrange>
      <Skills>large-scale ETL, columnar storage formats, query engines, high-volume time series data, data cataloging, lineage, metadata management systems, ML experiment tracking, Spark, BigQuery, DuckDB, Parquet</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Anthropic</Employername>
      <Employerlogo>https://logos.yubhub.co/anthropic.com.png</Employerlogo>
      <Employerdescription>Anthropic is a public benefit corporation that creates reliable, interpretable, and steerable AI systems.</Employerdescription>
      <Employerwebsite>https://www.anthropic.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/anthropic/jobs/5191226008</Applyto>
      <Location>San Francisco, CA | New York City, NY</Location>
      <Country></Country>
      <Postedate>2026-04-18</Postedate>
    </job>
    <job>
      <externalid>4075c787-328</externalid>
      <Title>Member of Technical Staff - Large Scale Data Infrastructure</Title>
      <Description><![CDATA[<p>We&#39;re looking for infrastructure engineers to work at peta-to-exabyte scale. You&#39;ll build data systems behind the largest training runs on thousands of GPUs, where fixing one bottleneck lets researchers train the next breakthrough model.</p>
<p><strong>What You&#39;ll Work On:</strong></p>
<ul>
<li>Scalable data loaders for training runs across thousands of GPUs</li>
<li>Efficient storage and retrieval systems for petabyte-scale datasets</li>
<li>Multi-cloud object storage abstraction</li>
<li>Execute large-scale data migrations across storage systems and providers</li>
<li>Debug and resolve performance bottlenecks in distributed data loading</li>
</ul>
<p><strong>Technical Focus:</strong></p>
<ul>
<li>Python, PyTorch DataLoader internals</li>
<li>Object storage (e.g. S3, Azure Blob, GCS)</li>
<li>Parquet for metadata</li>
<li>Video: ffmpeg, PyAV, codec fundamentals</li>
</ul>
<p><strong>What We&#39;re Looking For:</strong></p>
<ul>
<li>Built and operated data pipelines at petabyte scale</li>
<li>Optimized data loading</li>
<li>Worked with petabyte-scale video and image datasets</li>
<li>Written processing jobs operating on millions of files</li>
<li>Debugged distributed system bottlenecks across large fleets of machines</li>
</ul>
<p><strong>Nice to Have:</strong></p>
<ul>
<li>Experience streaming dataset formats (e.g. WebDataset)</li>
<li>Video codec internals and frame-accurate seeking</li>
<li>Distributed systems experience</li>
<li>Slurm and Kubernetes for job orchestration</li>
<li>Experience with object storage performance tuning across providers</li>
</ul>
<p><strong>How We Work Together:</strong></p>
<ul>
<li>We&#39;re a distributed team with real offices that people actually use. Depending on your role, you&#39;ll either join us in Freiburg or SF at least 2 days a week (or one full week every other week), or work remotely with a monthly in-person week to stay connected. We&#39;ll cover reasonable travel costs to make this possible. We think in-person time matters, and we&#39;ve structured things to make it accessible to all. We&#39;ll discuss what this will look like for the role during our interview process.</li>
</ul>
<p><strong>Everything we do is grounded in four values:</strong></p>
<ul>
<li>Obsessed. We are a frontier research lab. The science has to be right, the understanding deep, the product beautiful.</li>
<li>Low Ego. The work speaks. The best idea wins, no matter who said it. Credit is shared. Nobody is above any task.</li>
<li>Bold. We take the ambitious bet. We ship, we do not wait for conditions to be perfect.</li>
<li>Kind. People over politics. We treat each other with genuine warmth. Agency without empathy creates chaos.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>staff</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange>$180,000–$300,000 USD + Equity</Salaryrange>
      <Skills>Python, PyTorch, Data Loader Internals, Object Storage, Parquet, Video, ffmpeg, PyAV, Codec Fundamentals, WebDataset, Distributed Systems, Slurm, Kubernetes, Object Storage Performance Tuning</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Black Forest Labs</Employername>
      <Employerlogo>https://logos.yubhub.co/blackforestlabs.com.png</Employerlogo>
      <Employerdescription>Black Forest Labs is a research lab developing foundational technologies for generative models that power image and video creation.</Employerdescription>
      <Employerwebsite>https://www.blackforestlabs.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/blackforestlabs/jobs/5019171008</Applyto>
      <Location>Freiburg (Germany), San Francisco (USA)</Location>
      <Country></Country>
      <Postedate>2026-04-17</Postedate>
    </job>
    <job>
      <externalid>1739131f-5ae</externalid>
      <Title>Software Engineer - Silicon Validation Tools (SerDes)</Title>
      <Description><![CDATA[<p>Engineer the Future with Us</p>
<p>We currently have 614 open roles</p>
<p>Innovation Starts Here</p>
<p>Find Jobs For</p>
<p>Where?When autocomplete results are available use up and down arrows to review and enter to select. Touch device users, explore by touch or with swipe gestures.</p>
<p><strong>Software Engineer  - Silicon Validation Tools (SerDes)</strong></p>
<p>Mississauga, Ontario, Canada</p>
<p>Save</p>
<p>Category: EngineeringHire Type: Employee</p>
<p><strong>Job ID</strong> 16183<strong>Date posted</strong> 03/22/2026</p>
<p><strong><strong>We Are:</strong></strong></p>
<p>At Synopsys, we drive the innovations that shape the way we live and connect. Our technology is central to the Era of Pervasive Intelligence, from self-driving cars to learning machines. We lead in chip design, verification, and IP integration, empowering the creation of high-performance silicon chips and software content. Join us to transform the future through continuous technological innovation.</p>
<p><strong><strong>You Are:</strong></strong></p>
<p>You are a hands-on software engineer passionate about bridging the gap between hardware and software in high-speed silicon validation environments. You thrive in lab-heavy settings, working directly with advanced instruments and hardware abstraction layers. Your expertise in Python allows you to architect robust automation frameworks and reusable libraries, streamlining complex workflows and ensuring repeatable, high-quality results. You’re comfortable translating intricate MATLAB algorithms into efficient Python code, maintaining numerical equivalence and clear documentation for future maintainability. Your experience integrating C/C++ SDKs, DLLs, and libraries into Python empowers seamless cross-team collaboration with firmware and software teams worldwide. You are meticulous in your approach to code governance, CI/CD pipelines, and release management, ensuring that your code is reliable, maintainable, and well-documented.</p>
<p>You communicate clearly, translating technical requirements into actionable solutions and collaborating across validation, SDK, firmware, and software teams. You are proactive in identifying opportunities for process improvement, championing best practices, and mentoring teammates. Your curiosity drives you to stay current with industry trends, and your adaptability shines in fast-paced environments where innovation and continuous improvement are valued. You are committed to delivering intuitive GUIs and data tools that empower your team to make informed decisions, reduce operator error, and streamline bench operations. With a strong foundation in high-speed SerDes technologies and lab instrumentation, you are ready to make a lasting impact on industry-leading IP and silicon validation processes.</p>
<p><strong><strong>What You’ll Be Doing:</strong></strong></p>
<ul>
<li>Defining and maintaining Python automation architecture, folder/repo structure, and coding standards for lab environments.</li>
</ul>
<ul>
<li>Building hardware-abstraction layers for lab instruments (J-BERT, oscilloscopes, pattern generators, power supplies) using VISA/SCPI and vendor APIs.</li>
</ul>
<ul>
<li>Creating reusable libraries for test sequencing, calibration/adaptation flows, results logging, fault handling, and multi-bench resource scheduling.</li>
</ul>
<ul>
<li>Translating MATLAB algorithms and calibration scripts into robust Python (NumPy/SciPy/Pandas), ensuring numerical equivalence and documenting migration notes.</li>
</ul>
<ul>
<li>Developing and hardening drivers/wrappers for Keysight/Teledyne-LeCroy/R&amp;S instruments, including connection management, waveform acquisition, and compliance scripts.</li>
</ul>
<ul>
<li>Integrating SDK C/C++ code, DLLs, and shared libraries into Python via ctypes/cffi/SWIG, collaborating with firmware teams to validate features and APIs.</li>
</ul>
<ul>
<li>Managing code repositories (Git/Perforce), branching strategies, code reviews, release tagging, and CI/CD pipelines for linting, testing, and packaging.</li>
</ul>
<ul>
<li>Building internal GUIs for bench control, run setup, live plots, and progress tracking, as well as data processing pipelines for quick-turn analysis and report generation.</li>
</ul>
<p><strong><strong>The Impact You Will Have:</strong></strong></p>
<ul>
<li>Accelerate SerDes bring-up and characterization through standardized Python automation and instrument libraries, reducing time-to-first-measurement and increasing bench throughput.</li>
</ul>
<ul>
<li>Enhance efficiency and reliability of Silicon Validation processes with robust frameworks, CI-verified packages, and repeatable workflows.</li>
</ul>
<ul>
<li>Improve test quality, reproducibility, and data analysis by shipping versioned tools, enforcing parity on MATLAB to Python conversions, and delivering clear analysis artifacts.</li>
</ul>
<ul>
<li>Contribute directly to the success of high-speed, mixed-signal SerDes IP, enabling faster SDK/FW feature validation and higher confidence in release readiness.</li>
</ul>
<ul>
<li>Support cutting-edge IP solutions that drive industry innovation by turning validation requirements into dependable software and GUIs adopted across benches.</li>
</ul>
<ul>
<li>Streamline communication and day-to-day operations across Validation, FW, SDK, Software via documented interfaces, shared roadmaps, and predictable release notes.</li>
</ul>
<ul>
<li>Increase code reuse and maintainability with a clear repo structure, a reusable hardware-abstraction layer, and shared GUI components that standardize user experience.</li>
</ul>
<ul>
<li>Deliver intuitive, maintainable GUIs for bring-up and mass-char workflows, reducing operator error and enabling telemetry-driven improvements over time.</li>
</ul>
<ul>
<li>Foster a culture of innovation and continuous improvement through mentorship, code reviews, documentation, and knowledge sharing within the validation team.</li>
</ul>
<p><strong><strong>What You’ll Need:</strong></strong></p>
<ul>
<li>2–6 years professional software engineering experience building automation frameworks or tooling in a hardware or lab environment.</li>
</ul>
<ul>
<li>Strong proficiency in Python (OOP, packaging, virtual environments, logging, multithreading for instrument I/O).</li>
</ul>
<ul>
<li>Experience controlling lab instruments via SCPI/VISA and vendor SDKs/APIs; comfortable with Windows and Linux benches.</li>
</ul>
<ul>
<li>Proven skill integrating native libraries (C/C++ DLL/.so) into Python and debugging across the boundary.</li>
</ul>
<ul>
<li>Solid practice with Git: code reviews, branching strategies, conflict resolution, and release tagging.</li>
</ul>
<ul>
<li>Data skills: NumPy/Pandas, CSV/Parquet, plotting, and producing analysis artifacts usable by the team.</li>
</ul>
<ul>
<li>Clear communicator who can translate validation requirements into robust software and collaborate across SDK, FW, and SW teams.</li>
</ul>
<p><strong><strong>Who You Are:</strong></strong></p>
<ul>
<li>Analytical thinker with strong problem-solving skills.</li>
</ul>
<ul>
<li>Collaborative team player who thrives in cross-functional environments.</li>
</ul>
<ul>
<li>Detail-oriented and organized in both code and documentation.</li>
</ul>
<ul>
<li>Curious and eager to learn new technologies and methodologies.</li>
</ul>
<ul>
<li>Adaptable, proactive, and comfortable with ambiguity and innovation.</li>
</ul>
<ul>
<li>Effective communicator, able to translate complex requirements into actionable solutions.</li>
</ul>
<ul>
<li>Mentor and knowledge sharer, fostering a culture of continuous improvement.</li>
</ul>
<p><strong><strong>The Team You’ll Be A Part Of:</strong></strong></p>
<p>You’ll join the SerDes Silicon Validation team, a dynamic group of engineers dedicated to enabling high-speed IP innovation through rigorous hardware validation and software automation. The team works closely with worldwide software, firmware, and SDK teams, collaborating to deliver robust tools, frameworks, and GUIs that accelerate bring-up, characterization, and validation workflows. Together, you’ll drive advancements in silicon validation, empower efficient lab operations, and shape the future of high-performance IP solutions.</p>
<p><strong><strong>Rewards and Benefits:</strong></strong></p>
<p>We offer a comprehensive range of health, wellness, and financial benefits to cater to your needs. Our total rewards include both monetary and non-monetary offerings. Your recruiter will provide more details about the salary range and benefits during the hiring process.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, MATLAB, C/C++, SCPI/VISA, Git, NumPy, Pandas, CSV/Parquet, plotting</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Synopsys</Employername>
      <Employerlogo>https://logos.yubhub.co/careers.synopsys.com.png</Employerlogo>
      <Employerdescription>Synopsys is a leading provider of electronic design automation (EDA) software and intellectual property (IP) used in the design, verification, and manufacturing of electronic systems.</Employerdescription>
      <Employerwebsite>https://careers.synopsys.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://careers.synopsys.com/job/mississauga/software-engineer-silicon-validation-tools-serdes/44408/93120696128</Applyto>
      <Location>Mississauga</Location>
      <Country></Country>
      <Postedate>2026-04-05</Postedate>
    </job>
    <job>
      <externalid>d48b0655-2fa</externalid>
      <Title>Data/Infrastructure Advocate Engineer</Title>
      <Description><![CDATA[<p>At Hugging Face, we&#39;re on a journey to democratise good AI. As our first Data/Infrastructure Advocate Engineer, you&#39;ll bridge the gap between cutting-edge data infrastructure and the global community of data engineers, researchers, and developers.</p>
<p>You&#39;ll champion Xet storage on the Hugging Face Hub, empowering users to efficiently store, version, and collaborate on large-scale datasets. This role is for someone who thrives at the intersection of technical depth (storage, Parquet, deduplication) and community advocacy—helping define the future of open data workflows.</p>
<p>Your main missions will be:</p>
<ul>
<li>Grow and nurture the open-source data/infra community—launch initiatives, collaborate with data-focused groups, and organise events or challenges.</li>
<li>Promote the Hugging Face Hub as the go-to platform for data storage, versioning, and collaboration—curate and showcase datasets, benchmarks, and tools like Xet.</li>
<li>Highlight use cases like efficient large dataset updates, Parquet editing, and deduplication to demonstrate the Hub&#39;s value for data workflows.</li>
<li>Create demos, benchmarks, and tools (e.g., Colab notebooks) to illustrate best practices for data storage and versioning.</li>
<li>Experiment with Xet, Parquet, and other data formats to showcase their potential for ML and data engineering.</li>
<li>Produce high-quality tutorials, blog posts, and videos that make complex topics accessible.</li>
<li>Share insights on storage optimisation, dataset versioning, and deduplication to empower developers.</li>
<li>Actively participate in online communities (Discord, GitHub, forums) to highlight contributions, answer questions, and foster collaboration.</li>
<li>Ensure datasets and tools released on the Hub are well-documented, with clear examples, benchmarks, and use cases.</li>
</ul>
<p><strong>About you</strong></p>
<p>You&#39;re a great fit if you:</p>
<ul>
<li>Have strong technical skills in Python, data libraries (e.g., pandas, pyarrow, huggingface/datasets), and storage systems (Parquet, Open Table Formats, S3).</li>
<li>Are a hands-on builder who loves experimenting with data tools, storage optimisation, and dataset versioning.</li>
<li>Can clearly explain complex topics (e.g., deduplication, compression, Parquet editing) through writing, demos, or talks.</li>
<li>Are active in developer communities (GitHub, Discord, forums) and passionate about open source and knowledge sharing.</li>
<li>Thrive in fast-moving environments and enjoy building in public to inspire others.</li>
</ul>
<p>If you&#39;re interested in joining us but don&#39;t tick every box above, we still encourage you to apply! We&#39;re building a diverse team whose skills, experiences, and backgrounds complement one another.</p>
<p><strong>More about Hugging Face</strong></p>
<p>We are actively working to build a culture that values diversity, equity, and inclusivity. We are intentionally building a workplace where you feel respected and supported—regardless of who you are or where you come from.</p>
<p>Hugging Face is an equal opportunity employer, and we do not discriminate based on race, ethnicity, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or ability status.</p>
<p>We value development. You will work with some of the smartest people in our industry.</p>
<p>We provide all employees with reimbursement for relevant conferences, training, and education.</p>
<p>We care about your well-being. We offer flexible working hours and remote options.</p>
<p>We offer health, dental, and vision benefits for employees and their dependents.</p>
<p>We also offer parental leave and flexible paid time off.</p>
<p>We support our employees wherever they are. While we have office spaces in NYC and Paris, we&#39;re very distributed, and all remote employees have the opportunity to visit our offices.</p>
<p>If needed, we&#39;ll also outfit your workstation to ensure you succeed.</p>
<p>We want our teammates to be shareholders. All employees have company equity as part of their compensation package.</p>
<p>If we succeed in becoming a category-defining platform in machine learning and artificial intelligence, everyone enjoys the upside.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>entry</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, data libraries, pandas, pyarrow, huggingface/datasets, storage systems, Parquet, Open Table Formats, S3</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Hugging Face</Employername>
      <Employerlogo></Employerlogo>
      <Employerdescription>Hugging Face is a platform for AI builders with over 5 million users and 100k organisations.</Employerdescription>
      <Employerwebsite>https://huggingface.co/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://apply.workable.com/j/5CA82A9A98</Applyto>
      <Location>New York</Location>
      <Country></Country>
      <Postedate>2026-03-10</Postedate>
    </job>
    <job>
      <externalid>f81a1dc8-ca4</externalid>
      <Title>Data/Infrastructure Advocate Engineer - EMEA Remote</Title>
      <Description><![CDATA[<p>At Hugging Face, we&#39;re on a journey to democratize good AI. We are building the fastest growing platform for AI builders with over 5 million users &amp; 100k organisations who collectively shared over 1M models, 300k datasets &amp; 300k apps. Our open-source libraries have more than 400k+ stars on Github.</p>
<p>As our first Data/Infrastructure Advocate Engineer, you&#39;ll bridge the gap between cutting-edge data infrastructure and the global community of data engineers, researchers, and developers. You&#39;ll champion Xet storage on the Hugging Face Hub, empowering users to efficiently store, version, and collaborate on large-scale datasets.</p>
<p>This role is for someone who thrives at the intersection of technical depth (storage, Parquet, deduplication) and community advocacy—helping define the future of open data workflows. You&#39;ll collaborate with teams like Datasets, Hub, and Infrastructure to shape how developers interact with data on our platform, and inspire a community to build better, faster, and more scalable data pipelines.</p>
<p>Your Main Missions:</p>
<ul>
<li>Grow and nurture the open-source data/infra community—launch initiatives, collaborate with data-focused groups, and organise events or challenges. Engage with communities like Apache Parquet, Open Tables Formats, and data engineering forums to promote best practices and Hugging Face tools.</li>
</ul>
<ul>
<li>Promote the Hugging Face Hub as the go-to platform for data storage, versioning, and collaboration—curate and showcase datasets, benchmarks, and tools like Xet.</li>
</ul>
<ul>
<li>Highlight use cases like efficient large dataset updates, Parquet editing, and deduplication to demonstrate the Hub’s value for data workflows.</li>
</ul>
<ul>
<li>Create demos, benchmarks, and tools (e.g., Colab notebooks) to illustrate best practices for data storage and versioning.</li>
</ul>
<ul>
<li>Experiment with Xet, Parquet, and other data formats to showcase their potential for ML and data engineering.</li>
</ul>
<ul>
<li>Produce high-quality tutorials, blog posts, and videos that make complex topics accessible.</li>
</ul>
<ul>
<li>Share insights on storage optimisation, dataset versioning, and deduplication to empower developers.</li>
</ul>
<ul>
<li>Actively participate in online communities (Discord, GitHub, forums) to highlight contributions, answer questions, and foster collaboration.</li>
</ul>
<ul>
<li>Ensure datasets and tools released on the Hub are well-documented, with clear examples, benchmarks, and use cases.</li>
</ul>
<p><strong>About you</strong></p>
<p>You’re a great fit if you:</p>
<ul>
<li>Have strong technical skills in Python, data libraries (e.g., pandas, pyarrow, huggingface/datasets), and storage systems (Parquet, Open Table Formats, S3).</li>
</ul>
<ul>
<li>Are a hands-on builder who loves experimenting with data tools, storage optimisation, and dataset versioning.</li>
</ul>
<ul>
<li>Can clearly explain complex topics (e.g., deduplication, compression, Parquet editing) through writing, demos, or talks.</li>
</ul>
<ul>
<li>Are active in developer communities (GitHub, Discord, forums) and passionate about open source and knowledge sharing.</li>
</ul>
<ul>
<li>Thrive in fast-moving environments and enjoy building in public to inspire others.</li>
</ul>
<p>If you&#39;re interested in joining us but don&#39;t tick every box above, we still encourage you to apply! We&#39;re building a diverse team whose skills, experiences, and backgrounds complement one another. We&#39;re happy to consider where you might be able to make the biggest impact.</p>
<p><strong>More about Hugging Face</strong></p>
<p>We are actively working to build a culture that values diversity, equity, and inclusivity. We are intentionally building a workplace where you feel respected and supported—regardless of who you are or where you come from. We believe this is foundational to building a great company and community, as well as the future of machine learning more broadly. Hugging Face is an equal opportunity employer, and we do not discriminate based on race, ethnicity, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or ability status.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>entry</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, data libraries, pandas, pyarrow, huggingface/datasets, storage systems, Parquet, Open Table Formats, S3</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Hugging Face</Employername>
      <Employerlogo></Employerlogo>
      <Employerdescription>Hugging Face is a platform for AI builders with over 5 million users and 100k organisations.</Employerdescription>
      <Employerwebsite>https://huggingface.co/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://apply.workable.com/j/7C7F63E87A</Applyto>
      <Location>Paris</Location>
      <Country></Country>
      <Postedate>2026-03-10</Postedate>
    </job>
    <job>
      <externalid>0841fcf4-9ab</externalid>
      <Title>Data Engineer SE - II</Title>
      <Description><![CDATA[<p>We are on a mission to rid the world of bad customer service by “mobilizing” the way help is delivered. Today’s consumers want an always-available customer service experience that leaves them feeling valued and respected.</p>
<p>Helpshift helps B2B brands deliver this modern customer service experience through a mobile-first approach. We have changed how conversations take place, moving the conversation away from a slow, outdated email and desktop experience to an in-app chat experience that allows users to interact with brands in their own time.</p>
<p>Through our market-leading AI-powered chatbots and automation, we help brands deliver instant and rapid resolutions. Because agents play a key role in delivering help, our platform gives agents superpowers with automation and AI that simply works.</p>
<p><strong>About the Team</strong></p>
<p>Consumers care first and foremost about having their time valued by brands. Brands need insights into their customer service operation to serve their consumers effectively. Such insights and analytics are delivered through various data products like in-app analytics dashboards and data-sharing integrations.</p>
<p>The data platform team is responsible for designing, building, and maintaining the data infrastructure that enables such data and analytics products at scale. We build and manage data pipelines, databases, and other data structures to ensure that the data is reliable, accurate, and easily accessible.</p>
<p>We also enable internal stakeholders with business intelligence and machine learning teams with data ops. This team manages the platform that handles 2 Million events per minute and processes 1+ terabytes of data daily.</p>
<p><strong>About the Role</strong></p>
<ul>
<li>Building maintainable data pipelines both for data ingestion and operational analytics for data collected from 2 billion devices and 900M Monthly active users</li>
<li>Building customer-facing analytics products that deliver actionable insights and data, easily detect anomalies</li>
<li>Collaborating with data stakeholders to see what their data needs are and being a part of the analysis process</li>
<li>Write design specifications, test, deployment, and scaling plans for the data pipelines</li>
<li>Mentor people in the team &amp; organization</li>
</ul>
<p><strong>Requirements</strong></p>
<ul>
<li>3+ years of experience in building and running data pipelines that scale for TBs of data</li>
<li>Proficiency in high-level object-oriented programming language (Python or Java) is must</li>
<li>Experience in Cloud data platforms like Snowflake and AWS, EMR/Athena is a must</li>
<li>Experience in building modern data lakehouse architectures using Snowflake and columnar formats like Apache Iceberg/Hudi, Parquet, etc</li>
<li>Proficiency in Data modeling, SQL query profiling, and data warehousing skills is a must</li>
<li>Experience in distributed data processing engines like Apache Spark, Apache Flink, Datalfow/Apache Beam, etc</li>
<li>Knowledge of workflow orchestrators like Airflow, Dasgter, etc is a plus</li>
<li>Data visualization skills are a plus (PowerBI, Metabase, Tableau, Hex, Sigma, etc)</li>
<li>Excellent verbal and written communication skills</li>
<li>Bachelor’s Degree in Computer Science (or equivalent)</li>
</ul>
<p><strong>Benefits</strong></p>
<ul>
<li>Hybrid setup</li>
<li>Worker&#39;s insurance</li>
<li>Paid Time Offs</li>
<li>Other employee benefits to be discussed by our Talent Acquisition team in India.</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>Python, Java, Snowflake, AWS, EMR/Athena, Apache Iceberg/Hudi, Parquet, Apache Spark, Apache Flink, Datalflow/Apache Beam, Airflow, Data modeling, SQL query profiling, data warehousing, PowerBI, Metabase, Tableau, Hex, Sigma</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Helpshift</Employername>
      <Employerlogo>https://logos.yubhub.co/j.com.png</Employerlogo>
      <Employerdescription>Helpshift is a company that provides a mobile-first customer service experience for B2B brands. It has over 900 million active monthly consumers and is used by hundreds of leading brands.</Employerdescription>
      <Employerwebsite>https://apply.workable.com</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://apply.workable.com/j/D451DB2325</Applyto>
      <Location>Pune, Maharashtra, India</Location>
      <Country></Country>
      <Postedate>2026-03-09</Postedate>
    </job>
  </jobs>
</source>