{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/arrow"},"x-facet":{"type":"skill","slug":"arrow","display":"Arrow","count":7},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0987988a-011"},"title":"Feature Framework Engineer","description":"<p>The Systematic Platform Execution &amp; Exchange Data (SPEED) Team is at the core of Millennium&#39;s Equities, Quant Strategies, and Shared Services Technology organisation.</p>\n<p>We are looking for a C++ engineer to design and build high-performance, low-latency applications that process large volumes of real-time data.</p>\n<p>Principal Responsibilities:</p>\n<ul>\n<li>Design, implement, and maintain high-performance C++ services handling high message rates and low-latency workloads.</li>\n</ul>\n<ul>\n<li>Optimise existing components for latency, throughput, and CPU/memory efficiency.</li>\n</ul>\n<ul>\n<li>Develop and tune networking, messaging, and I/O layers to handle large data volumes reliably.</li>\n</ul>\n<ul>\n<li>Profile and debug performance issues at application, OS, and network levels.</li>\n</ul>\n<ul>\n<li>Collaborate with quantitative, trading, and infrastructure teams to translate requirements into robust technical solutions.</li>\n</ul>\n<ul>\n<li>Write clean, production-quality code with appropriate tests and documentation.</li>\n</ul>\n<ul>\n<li>Participate in code reviews, design discussions, and continuous improvement of engineering practices.</li>\n</ul>\n<p>Required Qualifications:</p>\n<ul>\n<li>Strong proficiency in modern C++ (C++17/20 or later).</li>\n</ul>\n<ul>\n<li>5+ years of experience.</li>\n</ul>\n<ul>\n<li>Analytics Focus: KDB / Q Experience for large market data, modern data analysis with pytorch, pandas and modern tooling including Apache arrow.</li>\n</ul>\n<ul>\n<li>Familiar with basics statistics as applied to financial research.</li>\n</ul>\n<ul>\n<li>Proven experience building performance-critical, real-time, or low-latency systems.</li>\n</ul>\n<ul>\n<li>Strong knowledge of computer science fundamentals: data structures, algorithms, memory management, and optimisation.</li>\n</ul>\n<ul>\n<li>Experience using profiling, benchmarking, and performance analysis tools.</li>\n</ul>\n<ul>\n<li>Proficiency with version control (Git) and standard build systems.</li>\n</ul>\n<ul>\n<li>Excellent problem-solving skills and attention to detail.</li>\n</ul>\n<ul>\n<li>Strong interpersonal skills with a proven ability to navigate large organisations.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Experience with kernel bypass or user-space networking technologies.</li>\n</ul>\n<ul>\n<li>Familiarity with AI productivity enhancing coding tools.</li>\n</ul>\n<ul>\n<li>Experience in financial markets, market data distribution, order routing, or exchange connectivity.</li>\n</ul>\n<ul>\n<li>Experience with monitoring/telemetry for high-performance systems.</li>\n</ul>\n<ul>\n<li>Familiarity with scripting languages for tooling and automation.</li>\n</ul>\n<ul>\n<li>AI: Familiarity with AI productivity enhancing coding tools.</li>\n</ul>\n<p>Personal Attributes:</p>\n<ul>\n<li>Obsessed with performance, measurement, and data-driven optimisation.</li>\n</ul>\n<ul>\n<li>Comfortable owning features end-to-end and operating in a production environment.</li>\n</ul>\n<ul>\n<li>Clear communicator who can work closely with both technical and non-technical stakeholders.</li>\n</ul>\n<ul>\n<li>Proactive, self-directed, and able to thrive in a highly iterative environment.</li>\n</ul>\n<p>The estimated base salary range for this position is $175,000 to $250,000, which is specific to New York and may change in the future.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0987988a-011","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Unknown","sameAs":"https://mlp.eightfold.ai","logo":"https://logos.yubhub.co/mlp.eightfold.ai.png"},"x-apply-url":"https://mlp.eightfold.ai/careers/job/755955682418","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$175,000 to $250,000","x-skills-required":["modern C++","KDB / Q","pytorch","pandas","Apache arrow","data structures","algorithms","memory management","optimisation","profiling","benchmarking","performance analysis tools","version control","standard build systems"],"x-skills-preferred":["kernel bypass","user-space networking technologies","AI productivity enhancing coding tools","financial markets","market data distribution","order routing","exchange connectivity","monitoring/telemetry","scripting languages"],"datePosted":"2026-04-18T22:14:03.382Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York, New York, United States of America"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"modern C++, KDB / Q, pytorch, pandas, Apache arrow, data structures, algorithms, memory management, optimisation, profiling, benchmarking, performance analysis tools, version control, standard build systems, kernel bypass, user-space networking technologies, AI productivity enhancing coding tools, financial markets, market data distribution, order routing, exchange connectivity, monitoring/telemetry, scripting languages","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":175000,"maxValue":250000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d0ee3e8e-4f6"},"title":"Staff Software Engineer","description":"<p>About Us</p>\n<p>dbt Labs is the pioneer of analytics engineering, helping data teams transform raw data into reliable, actionable insights.</p>\n<p>As of February 2025, we&#39;ve surpassed $100 million in annual recurring revenue (ARR) and serve more than 5,400 dbt Platform customers, including AstraZeneca, Sky, Nasdaq, Volvo, JetBlue, and SafetyCulture.</p>\n<p>We&#39;re backed by top-tier investors including Andreessen Horowitz, Sequoia Capital, and Altimeter.</p>\n<p><strong>About The Team</strong></p>\n<p>dbt Fusion is building the next generation of data execution and connectivity infrastructure, enabling dbt workloads to run efficiently across diverse compute engines and data platforms.</p>\n<p>As a Senior Engineer on the Fusion Adapters and Connectivity team, you&#39;ll design and ship core abstractions powering how dbt communicates with execution systems , leveraging Rust, Go, Arrow, and emerging open standards.</p>\n<p>This is a rare opportunity to work at the intersection of systems programming, database internals, and high-visibility open-source development.</p>\n<p>Your work will shape a foundational platform leveraged across the dbt ecosystem and the broader data community.</p>\n<p><strong>You are a good fit if you have:</strong></p>\n<ul>\n<li>Strong programming background in Rust, Go, C++ or similar performance-oriented languages.</li>\n</ul>\n<ul>\n<li>Experience designing or maintaining SDKs, libraries, connectors, or compute/data integration codebases.</li>\n</ul>\n<ul>\n<li>Exposure to data warehouses, query engines, Arrow/columnar ecosystems, or execution runtimes.</li>\n</ul>\n<ul>\n<li>A desire to build foundational platform components that other teams and community members rely on.</li>\n</ul>\n<ul>\n<li>Comfort working in public code review loops, async-first communication, and collaborative RFC processes.</li>\n</ul>\n<ul>\n<li>A mindset grounded in debuggability, reliability, and ownership in ambiguous problem spaces.</li>\n</ul>\n<p><strong>In this role, you can expect to:</strong></p>\n<ul>\n<li>Design, build, and maintain Rust-first connectivity layers, execution APIs, and adapter scaffolding.</li>\n</ul>\n<ul>\n<li>Partner with teams building the dbt compiler, semantic layer, and runtime to evolve adapter interfaces and system boundaries.</li>\n</ul>\n<ul>\n<li>Contribute to Arrow/ADBC and other open-source specifications or implementations, strengthening the data ecosystem.</li>\n</ul>\n<ul>\n<li>Own CI, testing frameworks, profiling, error reporting surfaces, and release readiness for Fusion adapters.</li>\n</ul>\n<ul>\n<li>Debug complex interoperability and performance issues across drivers, engines, and compute domains.</li>\n</ul>\n<ul>\n<li>Collaborate with internal and community maintainers to review PRs, write RFCs, and evolve public code architectures.</li>\n</ul>\n<ul>\n<li>Mentor engineers on systems best practices and contribute to shared patterns around resilience, debuggability, and API clarity.</li>\n</ul>\n<p><strong>You&#39;ll have an edge if you have:</strong></p>\n<ul>\n<li>Contributed to or interacted with Arrow, ADBC, DuckDB, Presto, DataFusion, Spark, ClickHouse, or similar engines.</li>\n</ul>\n<ul>\n<li>Experience shaping adapter/plugin standards, driver contracts, or architectural interfaces used by others.</li>\n</ul>\n<ul>\n<li>Familiarity with Rust async ecosystems (tokio, tower, tracing) or Go concurrency practices.</li>\n</ul>\n<ul>\n<li>Prior OSS governance experience , triaging issues, reviewing PRs, or working with community maintainers.</li>\n</ul>\n<ul>\n<li>An interest in building developer-experience layers or scaffolding frameworks for adapter authors.</li>\n</ul>\n<p><strong>Qualifications:</strong></p>\n<ul>\n<li>6+ years experience in software engineering, with strong systems-level skills.</li>\n</ul>\n<ul>\n<li>2+ years working in open-source, SDK, runtime, or low-level integration environments.</li>\n</ul>\n<ul>\n<li>Bachelor&#39;s degree in Computer Science / related field or equivalent experience through industry OSS contributions.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d0ee3e8e-4f6","directApply":true,"hiringOrganization":{"@type":"Organization","name":"dbt Labs","sameAs":"https://www.getdbt.com/","logo":"https://logos.yubhub.co/getdbt.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dbtlabsinc/jobs/4641221005","x-work-arrangement":"remote","x-experience-level":"staff","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Rust","Go","C++","Arrow","ADBC","DuckDB","Presto","DataFusion","Spark","ClickHouse"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:52:31.073Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"India - Remote"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Rust, Go, C++, Arrow, ADBC, DuckDB, Presto, DataFusion, Spark, ClickHouse"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_72eaaa6e-3c0"},"title":"Founding Engineer - Reporting & Statements","description":"<p>Join us as a founding engineer on our Reporting &amp; Statements team. You&#39;ll design the systems that power every financial report and statement we deliver from monthly reports to daily statements to custom client requests. We&#39;re building automated frameworks that guarantee accuracy and consistency for every number we send to clients.</p>\n<p><strong>Technical Skills:</strong></p>\n<ul>\n<li>Evolve our architecture from decentralized reporting scripts to a centralized, framework-based delivery system</li>\n<li>Build automated validation and reconciliation that lets us scale without adding manual oversight</li>\n</ul>\n<p><strong>Complexity and Impact of Work:</strong></p>\n<ul>\n<li>Design data models that become a trusted, shared source of truth for downstream product teams and external APIs</li>\n<li>Navigate complexity across multiple product data streams, applying consistent logic to all financial statements</li>\n</ul>\n<p><strong>Organizational Knowledge:</strong></p>\n<ul>\n<li>Work with Product and Foundations teams to standardize how we capture and represent financial data</li>\n<li>Create self-service frameworks so other teams can add new report types through configuration instead of code</li>\n</ul>\n<p><strong>Communication and Influence:</strong></p>\n<ul>\n<li>Listen to product stakeholders to stay ahead of scaling needs for client-facing data</li>\n<li>Help mature our engineering culture by advocating for and modeling &#39;Data as a Product&#39; principles and high-quality engineering standards</li>\n</ul>\n<p><strong>You may be a fit for this role if you:</strong></p>\n<ul>\n<li>7+ years building data systems: You have experience creating internal tools, frameworks, or engines that handle 10x scale</li>\n<li>Financial domain experience: You&#39;ve worked in fintech, banking, or other environments where numbers matter. You understand what a &#39;Statement of Record&#39; means and the precision it demands.</li>\n<li>Systems thinking: You consider the next 100 products, not just the current one. You value extensible systems over one-off pipelines.</li>\n<li>Solid technical foundation: You&#39;re proficient with Python (Pandas/Polars/Arrow) and SQL, with experience in BigQuery or similar cloud warehouses and modern orchestration tools like Airflow or Dagster.</li>\n</ul>\n<p><strong>Although not a requirement, bonus points if:</strong></p>\n<ul>\n<li>You&#39;ve been a data consumer: Prior experience as a financial or business analyst gives you the perspective to design truly usable data models.</li>\n<li>You care about performance: You enjoy making data move faster and cheaper, whether through ADBC, multiprocessing, vectorized operations, or other optimizations.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_72eaaa6e-3c0","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anchorage Digital","sameAs":"https://www.anchorage.co/","logo":"https://logos.yubhub.co/anchorage.co.png"},"x-apply-url":"https://jobs.lever.co/anchorage/5bcfc8f2-5f26-4f72-8ca7-f4b20ee7f7db","x-work-arrangement":"remote","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","SQL","BigQuery","Airflow","Dagster","Pandas","Polars","Arrow"],"x-skills-preferred":[],"datePosted":"2026-04-17T12:23:54.465Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York City"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, SQL, BigQuery, Airflow, Dagster, Pandas, Polars, Arrow"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d48b0655-2fa"},"title":"Data/Infrastructure Advocate Engineer","description":"<p>At Hugging Face, we&#39;re on a journey to democratise good AI. As our first Data/Infrastructure Advocate Engineer, you&#39;ll bridge the gap between cutting-edge data infrastructure and the global community of data engineers, researchers, and developers.</p>\n<p>You&#39;ll champion Xet storage on the Hugging Face Hub, empowering users to efficiently store, version, and collaborate on large-scale datasets. This role is for someone who thrives at the intersection of technical depth (storage, Parquet, deduplication) and community advocacy—helping define the future of open data workflows.</p>\n<p>Your main missions will be:</p>\n<ul>\n<li>Grow and nurture the open-source data/infra community—launch initiatives, collaborate with data-focused groups, and organise events or challenges.</li>\n<li>Promote the Hugging Face Hub as the go-to platform for data storage, versioning, and collaboration—curate and showcase datasets, benchmarks, and tools like Xet.</li>\n<li>Highlight use cases like efficient large dataset updates, Parquet editing, and deduplication to demonstrate the Hub&#39;s value for data workflows.</li>\n<li>Create demos, benchmarks, and tools (e.g., Colab notebooks) to illustrate best practices for data storage and versioning.</li>\n<li>Experiment with Xet, Parquet, and other data formats to showcase their potential for ML and data engineering.</li>\n<li>Produce high-quality tutorials, blog posts, and videos that make complex topics accessible.</li>\n<li>Share insights on storage optimisation, dataset versioning, and deduplication to empower developers.</li>\n<li>Actively participate in online communities (Discord, GitHub, forums) to highlight contributions, answer questions, and foster collaboration.</li>\n<li>Ensure datasets and tools released on the Hub are well-documented, with clear examples, benchmarks, and use cases.</li>\n</ul>\n<p><strong>About you</strong></p>\n<p>You&#39;re a great fit if you:</p>\n<ul>\n<li>Have strong technical skills in Python, data libraries (e.g., pandas, pyarrow, huggingface/datasets), and storage systems (Parquet, Open Table Formats, S3).</li>\n<li>Are a hands-on builder who loves experimenting with data tools, storage optimisation, and dataset versioning.</li>\n<li>Can clearly explain complex topics (e.g., deduplication, compression, Parquet editing) through writing, demos, or talks.</li>\n<li>Are active in developer communities (GitHub, Discord, forums) and passionate about open source and knowledge sharing.</li>\n<li>Thrive in fast-moving environments and enjoy building in public to inspire others.</li>\n</ul>\n<p>If you&#39;re interested in joining us but don&#39;t tick every box above, we still encourage you to apply! We&#39;re building a diverse team whose skills, experiences, and backgrounds complement one another.</p>\n<p><strong>More about Hugging Face</strong></p>\n<p>We are actively working to build a culture that values diversity, equity, and inclusivity. We are intentionally building a workplace where you feel respected and supported—regardless of who you are or where you come from.</p>\n<p>Hugging Face is an equal opportunity employer, and we do not discriminate based on race, ethnicity, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or ability status.</p>\n<p>We value development. You will work with some of the smartest people in our industry.</p>\n<p>We provide all employees with reimbursement for relevant conferences, training, and education.</p>\n<p>We care about your well-being. We offer flexible working hours and remote options.</p>\n<p>We offer health, dental, and vision benefits for employees and their dependents.</p>\n<p>We also offer parental leave and flexible paid time off.</p>\n<p>We support our employees wherever they are. While we have office spaces in NYC and Paris, we&#39;re very distributed, and all remote employees have the opportunity to visit our offices.</p>\n<p>If needed, we&#39;ll also outfit your workstation to ensure you succeed.</p>\n<p>We want our teammates to be shareholders. All employees have company equity as part of their compensation package.</p>\n<p>If we succeed in becoming a category-defining platform in machine learning and artificial intelligence, everyone enjoys the upside.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d48b0655-2fa","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Hugging Face","sameAs":"https://huggingface.co/"},"x-apply-url":"https://apply.workable.com/j/5CA82A9A98","x-work-arrangement":"remote","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","data libraries","pandas","pyarrow","huggingface/datasets","storage systems","Parquet","Open Table Formats","S3"],"x-skills-preferred":[],"datePosted":"2026-03-10T11:34:41.656Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"New York"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, data libraries, pandas, pyarrow, huggingface/datasets, storage systems, Parquet, Open Table Formats, S3"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_f81a1dc8-ca4"},"title":"Data/Infrastructure Advocate Engineer - EMEA Remote","description":"<p>At Hugging Face, we&#39;re on a journey to democratize good AI. We are building the fastest growing platform for AI builders with over 5 million users &amp; 100k organisations who collectively shared over 1M models, 300k datasets &amp; 300k apps. Our open-source libraries have more than 400k+ stars on Github.</p>\n<p>As our first Data/Infrastructure Advocate Engineer, you&#39;ll bridge the gap between cutting-edge data infrastructure and the global community of data engineers, researchers, and developers. You&#39;ll champion Xet storage on the Hugging Face Hub, empowering users to efficiently store, version, and collaborate on large-scale datasets.</p>\n<p>This role is for someone who thrives at the intersection of technical depth (storage, Parquet, deduplication) and community advocacy—helping define the future of open data workflows. You&#39;ll collaborate with teams like Datasets, Hub, and Infrastructure to shape how developers interact with data on our platform, and inspire a community to build better, faster, and more scalable data pipelines.</p>\n<p>Your Main Missions:</p>\n<ul>\n<li>Grow and nurture the open-source data/infra community—launch initiatives, collaborate with data-focused groups, and organise events or challenges. Engage with communities like Apache Parquet, Open Tables Formats, and data engineering forums to promote best practices and Hugging Face tools.</li>\n</ul>\n<ul>\n<li>Promote the Hugging Face Hub as the go-to platform for data storage, versioning, and collaboration—curate and showcase datasets, benchmarks, and tools like Xet.</li>\n</ul>\n<ul>\n<li>Highlight use cases like efficient large dataset updates, Parquet editing, and deduplication to demonstrate the Hub’s value for data workflows.</li>\n</ul>\n<ul>\n<li>Create demos, benchmarks, and tools (e.g., Colab notebooks) to illustrate best practices for data storage and versioning.</li>\n</ul>\n<ul>\n<li>Experiment with Xet, Parquet, and other data formats to showcase their potential for ML and data engineering.</li>\n</ul>\n<ul>\n<li>Produce high-quality tutorials, blog posts, and videos that make complex topics accessible.</li>\n</ul>\n<ul>\n<li>Share insights on storage optimisation, dataset versioning, and deduplication to empower developers.</li>\n</ul>\n<ul>\n<li>Actively participate in online communities (Discord, GitHub, forums) to highlight contributions, answer questions, and foster collaboration.</li>\n</ul>\n<ul>\n<li>Ensure datasets and tools released on the Hub are well-documented, with clear examples, benchmarks, and use cases.</li>\n</ul>\n<p><strong>About you</strong></p>\n<p>You’re a great fit if you:</p>\n<ul>\n<li>Have strong technical skills in Python, data libraries (e.g., pandas, pyarrow, huggingface/datasets), and storage systems (Parquet, Open Table Formats, S3).</li>\n</ul>\n<ul>\n<li>Are a hands-on builder who loves experimenting with data tools, storage optimisation, and dataset versioning.</li>\n</ul>\n<ul>\n<li>Can clearly explain complex topics (e.g., deduplication, compression, Parquet editing) through writing, demos, or talks.</li>\n</ul>\n<ul>\n<li>Are active in developer communities (GitHub, Discord, forums) and passionate about open source and knowledge sharing.</li>\n</ul>\n<ul>\n<li>Thrive in fast-moving environments and enjoy building in public to inspire others.</li>\n</ul>\n<p>If you&#39;re interested in joining us but don&#39;t tick every box above, we still encourage you to apply! We&#39;re building a diverse team whose skills, experiences, and backgrounds complement one another. We&#39;re happy to consider where you might be able to make the biggest impact.</p>\n<p><strong>More about Hugging Face</strong></p>\n<p>We are actively working to build a culture that values diversity, equity, and inclusivity. We are intentionally building a workplace where you feel respected and supported—regardless of who you are or where you come from. We believe this is foundational to building a great company and community, as well as the future of machine learning more broadly. Hugging Face is an equal opportunity employer, and we do not discriminate based on race, ethnicity, religion, colour, national origin, gender, sexual orientation, age, marital status, veteran status, or ability status.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_f81a1dc8-ca4","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Hugging Face","sameAs":"https://huggingface.co/"},"x-apply-url":"https://apply.workable.com/j/7C7F63E87A","x-work-arrangement":"remote","x-experience-level":"entry","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","data libraries","pandas","pyarrow","huggingface/datasets","storage systems","Parquet","Open Table Formats","S3"],"x-skills-preferred":[],"datePosted":"2026-03-10T11:34:10.184Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Paris"}},"jobLocationType":"TELECOMMUTE","employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, data libraries, pandas, pyarrow, huggingface/datasets, storage systems, Parquet, Open Table Formats, S3"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_d92bf714-0ed"},"title":"Python Financial Model Engineer, Associate","description":"<p>About this role</p>\n<p>We are looking for a self-motivated software engineer to onboard models to our platform. Collaborate with modellers on the implementation and deployment of financial risk models. Understand client requirements and translate them into software engineering tasks.</p>\n<p>Our team</p>\n<ul>\n<li>Is passionate about technology and solving complex problems.</li>\n<li>Develops in Python, working with technologies like Pandas, Apache Arrow, Snowflake, Prefect, Docker, and Azure DevOps.</li>\n<li>Consists of technologists that unlock constant innovation.</li>\n<li>Constantly challenges the technology status quo and looks for ways to improve the platform.</li>\n</ul>\n<p>Key Responsibilities</p>\n<p>We expect the role to involve the following core responsibilities and would expect a successful candidate to be able to demonstrate skills or experience with the following (not in order of priority):</p>\n<ul>\n<li>Quickly learn the platform and act as a subject matter expert towards modelling teams and product analysts.</li>\n<li>Work with modellers and product analysts to understand the business and their requirements. Help implement those on our platform using engineering best practices.</li>\n<li>Facilitate technical design and code review sessions to ensure software meets functional and compatibility requirements, as well as high quality standards.</li>\n<li>Stay abreast of the latest developments in machine learning, quantitative finance, and technology to incorporate innovative solutions into our platform.</li>\n<li>Enhance the performance of existing models, ensuring they operate efficiently at scale.</li>\n<li>Implementation and maintenance of a standard data / technology deployment workflow to ensure that all deliverables/enhancements are delivered in a disciplined and robust manner.</li>\n<li>Ensure operational readiness of the product and meet customer commitments with regards to incident SLAs.</li>\n</ul>\n<p>Skillset</p>\n<ul>\n<li>Strong experience (3+ years) in Python is crucial</li>\n<li>Bachelor’s (BSc) or higher degree in Computer Science or related field</li>\n<li>Experience with Pandas, Apache Arrow, Snowflake, (Prefect is a plus)</li>\n<li>Good understanding of Object-Oriented Design principles</li>\n<li>Fluency with AI coding tools and the use of LLM in everyday development</li>\n<li>Good understanding of fundamental Algorithms and Data Structures</li>\n<li>Knowledge of Azure DevOps and git, CI/CD</li>\n<li>Good understanding of unit tests, integration and regression tests, and their importance</li>\n<li>An aptitude for designing data models and pipelines is a plus</li>\n<li>Ability to understand advanced mathematical and statistical methods and concepts</li>\n<li>Fluency in reading, writing and speaking English</li>\n</ul>\n<p>Personal Qualities</p>\n<ul>\n<li>Team player</li>\n<li>Problem-solving skills</li>\n<li>Critical and analytical thinking</li>\n<li>Technical curiosity</li>\n<li>Adaptable</li>\n</ul>\n<p>Our benefits</p>\n<p>To help you stay energized, engaged and inspired, we offer a wide range of employee benefits including: retirement investment and tools designed to help you in building a sound financial future; access to education reimbursement; comprehensive resources to support your physical health and emotional well-being; family support programs; and Flexible Time Off (FTO) so you can relax, recharge and be there for the people you care about.</p>\n<p>Our hybrid work model</p>\n<p>BlackRock’s hybrid work model is designed to enable a culture of collaboration and apprenticeship that enriches the experience of our employees, while supporting flexibility for all. Employees are currently required to work at least 4 days in the office per week, with the flexibility to work from home 1 day a week. Some business groups may require more time in the office due to their roles and responsibilities. We remain focused on increasing the impactful moments that arise when we work together in person – aligned with our commitment to performance and innovation. As a new joiner, you can count on this hybrid model to accelerate your learning and onboarding experience here at BlackRock.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_d92bf714-0ed","directApply":true,"hiringOrganization":{"@type":"Organization","name":"BlackRock","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/fj1ayC3FJpaeEtw176Tee2/python-financial-model-engineer%2C-associate-in-budapest-at-blackrock","x-work-arrangement":"hybrid","x-experience-level":"mid","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Pandas","Apache Arrow","Snowflake","Prefect","Docker","Azure DevOps","Object-Oriented Design principles","AI coding tools","LLM","fundamental Algorithms and Data Structures","Azure DevOps and git","CI/CD","unit tests","integration and regression tests"],"x-skills-preferred":["advanced mathematical and statistical methods and concepts"],"datePosted":"2026-03-09T16:45:41.136Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Budapest, Hungary"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Pandas, Apache Arrow, Snowflake, Prefect, Docker, Azure DevOps, Object-Oriented Design principles, AI coding tools, LLM, fundamental Algorithms and Data Structures, Azure DevOps and git, CI/CD, unit tests, integration and regression tests, advanced mathematical and statistical methods and concepts"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_10d04dca-482"},"title":"Python Financial Engineering Platform Lead, Vice President","description":"<p>About this role</p>\n<p>BlackRock is seeking a self-motivated lead software engineer to spearhead the development of its Python financial engineering platform. The platform enables research, implementation, delivery, and execution of financial risk models for internal partners and external clients.</p>\n<p>What will you be doing?</p>\n<ul>\n<li>Lead platform development supporting both financial engineers and researchers.</li>\n<li>Facilitate technical design and code review sessions to ensure software meets functional and compatibility requirements, as well as high quality standards.</li>\n<li>Build widely used and reliable fundamental components as part of the platform, distributed as Python libraries.</li>\n<li>Stay abreast of the latest developments in machine learning, quantitative finance, and technology to incorporate innovative solutions into our platform.</li>\n</ul>\n<p>Key Responsibilities</p>\n<ul>\n<li>Quickly learn the platform and act as a subject matter expert towards modelling teams.</li>\n<li>Build high quality software that improves the user experience of the downstream modeller and developer.</li>\n<li>Enhance the performance of existing models, ensuring they operate efficiently at scale.</li>\n<li>Implementation and maintenance of a standard data / technology deployment workflow to ensure that all deliverables/enhancements are delivered in a disciplined and robust manner.</li>\n</ul>\n<p>Skillset</p>\n<ul>\n<li>Strong experience (5+ years) in Python is crucial.</li>\n<li>Bachelor&#39;s (BSc) or higher degree in Computer Science or equivalent field.</li>\n<li>Experience with Pandas, Apache Arrow, Snowflake, (Prefect is a plus).</li>\n<li>Good understanding of Object-Oriented Design principles.</li>\n<li>Good understanding of fundamental Algorithms and Data Structures.</li>\n<li>Knowledge of Azure DevOps and git, CI/CD.</li>\n<li>Good understanding of unit tests, integration and regression tests, and their importance.</li>\n</ul>\n<p>Personal Qualities</p>\n<ul>\n<li>Team player.</li>\n<li>Problem-solving skills.</li>\n<li>Critical and analytical thinking.</li>\n<li>Technical curiosity.</li>\n<li>Adaptable.</li>\n</ul>\n<p>Our benefits</p>\n<ul>\n<li>Retirement investment and tools designed to help you in building a sound financial future.</li>\n<li>Access to education reimbursement.</li>\n<li>Comprehensive resources to support your physical health and emotional well-being.</li>\n<li>Family support programs.</li>\n<li>Flexible Time Off (FTO) so you can relax, recharge and be there for the people you care about.</li>\n</ul>\n<p>Our hybrid work model</p>\n<p>BlackRock&#39;s hybrid work model is designed to enable a culture of collaboration and apprenticeship that enriches the experience of our employees, while supporting flexibility for all. Employees are currently required to work at least 4 days in the office per week, with the flexibility to work from home 1 day a week.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_10d04dca-482","directApply":true,"hiringOrganization":{"@type":"Organization","name":"BlackRock","sameAs":"https://jobs.workable.com","logo":"https://logos.yubhub.co/view.com.png"},"x-apply-url":"https://jobs.workable.com/view/2mb7rZ3FNbXbRf6dpxMmaf/python-financial-engineering-platform-lead%2C-vice-president-in-budapest-at-blackrock","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Python","Pandas","Apache Arrow","Snowflake","Azure DevOps","git","CI/CD","unit tests","integration and regression tests"],"x-skills-preferred":["Prefect"],"datePosted":"2026-03-09T16:42:39.478Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Budapest, Hungary"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Finance","skills":"Python, Pandas, Apache Arrow, Snowflake, Azure DevOps, git, CI/CD, unit tests, integration and regression tests, Prefect"}]}