<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>8e20eaf6-7f6</externalid>
      <Title>Data Operations, Associate</Title>
      <Description><![CDATA[<p>About this role</p>
<p>Own advanced operational support and stability for enterprise data platforms, acting as the primary L2/L3 interface for ETL/ELT pipelines, orchestration, observability, and Snowflake workloads. The role bridges execution and engineering, with accountability for incident resolution, platform reliability, and operational improvement.</p>
<p>Key Responsibilities</p>
<ul>
<li>Own L1/L2 operational support for production data platforms, including data lakes, streaming pipelines, and Snowflake-based analytics.</li>
<li>Diagnose and resolve complex failures in ETL/ELT pipelines and orchestration frameworks, partnering with engineering where required.</li>
<li>Actively manage incidents, including impact assessment, remediation coordination, and post incident documentation.</li>
<li>Improve monitoring, alerting, and observability coverage, identifying gaps and driving instrumentation enhancements.</li>
<li>Support onboarding of new pipelines and data products by validating operational readiness, scalability, and reliability.</li>
<li>Analyze recurring incidents and data quality issues, contributing to root cause analysis (RCA) and long-term remediation.</li>
<li>Mentor analysts through guidance on operational best practices, troubleshooting, and platform behavior.</li>
<li>Contribute to automation initiatives to reduce manual effort and improve operational efficiency.</li>
</ul>
<p>Our benefits</p>
<p>To help you stay energized, engaged and inspired, we offer a wide range of employee benefits including: retirement investment and tools designed to help you in building a sound financial future; access to education reimbursement; comprehensive resources to support your physical health and emotional well-being; family support programs; and Flexible Time Off (FTO) so you can relax, recharge and be there for the people you care about.</p>
<p>Our hybrid work model</p>
<p>BlackRock’s hybrid work model is designed to enable a culture of collaboration and apprenticeship that enriches the experience of our employees, while supporting flexibility for all. Employees are currently required to work at least 4 days in the office per week, with the flexibility to work from home 1 day a week. Some business groups may require more time in the office due to their roles and responsibilities.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>entry</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>enterprise data platforms, ETL/ELT pipelines, orchestration, observability, Snowflake workloads, AWS, Azure, GCP, cloud-native data services, monitoring, alerting, observability systems</Skills>
      <Category>Engineering</Category>
      <Industry>Finance</Industry>
      <Employername>BlackRock</Employername>
      <Employerlogo>https://logos.yubhub.co/blackrock.com.png</Employerlogo>
      <Employerdescription>BlackRock is a global investment management corporation that provides a range of investment management services to institutional and retail clients.</Employerdescription>
      <Employerwebsite>https://www.blackrock.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.workable.com/view/v9REx3w1EEK7y2df1zPkqK/data-operations%2C-associate-in-edinburgh-at-blackrock</Applyto>
      <Location>Edinburgh</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>7d57ab2d-f3b</externalid>
      <Title>Cloud Solution Architect</Title>
      <Description><![CDATA[<p>At Ford Motor Company, we believe freedom of movement drives human progress. We also believe in providing you with the freedom to define and realize your dreams. With our incredible plans for the future of mobility, we have a wide variety of opportunities for you to accelerate your career potential as you help us define tomorrow&#39;s transportation.</p>
<p>If you&#39;re looking for the chance to leverage advanced technology to redefine the transportation landscape, enhance the customer experience, and improve people&#39;s lives: this is the opportunity for you. Join us and challenge your IT expertise and analytical skills to help create vehicles that are as smart as you are.</p>
<p>To meet the growing needs of the Customer analytics business, the team is looking for a self-motivated, technically proficient individual to craft and shepherd coherent solutions. This will require collaboration with a range of stakeholders to clarify requirements, establish pragmatic approaches, and support and articulate decisions over time. You will join a cloud architecture team that works closely with engineering teams and other architects across the organisation.</p>
<p><strong>Responsibilities</strong></p>
<p><strong>Technical Requirements</strong></p>
<ul>
<li>Extensive experience with Google Cloud Platform (GCP), specifically BigQuery, Vertex AI, Dataflow, Dataproc, Cloud Run, CloudSQL, Spanner and Apigee.</li>
</ul>
<ul>
<li>Security &amp; Networking: Strong understanding of cloud security protocols, IAM, encryption, and complex network topologies.</li>
</ul>
<ul>
<li>Data Management: Proficiency in Enterprise Data Platforms, Data mesh architecture and data-driven architectural patterns.</li>
</ul>
<ul>
<li>DevOps Tooling: Hands-on experience with GitHub, SonarQube, Checkmarx, and FOSSA.</li>
</ul>
<ul>
<li>Software Engineering: Strong background in building Web Services and maintaining Clean Code standards.</li>
</ul>
<p><strong>Technical Leadership &amp; Strategy</strong></p>
<ul>
<li>System Design: Work with engineering teams to refine system designs, evangelising for horizontal scalability, resilience, and Clean Code compliance.</li>
</ul>
<ul>
<li>Product Collaboration: Partner with Product Managers to decompose complex business needs into incremental, production-ready user stories within an Agile/Sprint methodology.</li>
</ul>
<ul>
<li>Architectural Governance: Assess and document the rationale and tradeoffs for technical decisions; contribute to the broader Cloud Architecture team to improve global practices.</li>
</ul>
<ul>
<li>DevOps Excellence: Utilise and improve CI/CD pipelines using GitHub and automated testing/security tools to maximise deployment efficiency and minimise risk.</li>
</ul>
<p><strong>Cloud, Networking &amp; Security</strong></p>
<ul>
<li>Secure Infrastructure: Serve as the primary architect for cloud solutions, ensuring &#39;Secure-by-Design&#39; principles are applied across Google Cloud services (Dataflow, Dataproc, CloudRun, CloudSQL, Spanner).</li>
</ul>
<ul>
<li>Advanced Networking: Design and optimise cloud networking configurations, including VPCs, Service Controls, Load Balancing, and Private Service Connect to ensure high availability and low latency.</li>
</ul>
<ul>
<li>Cyber Security Oversight: Integrate security scanning and compliance into the architecture (utilising Checkmarx, SonarQube, and FOSSA). Proactively address vulnerabilities in distributed systems and AI models (e.g., OWASP Top 10 for LLMs).</li>
</ul>
<ul>
<li>API &amp; Data Contracts: Bolster &#39;Data as a Product&#39; practices by enforcing strict API standards and data contracts to ensure seamless, secure interoperability between services.</li>
</ul>
<ul>
<li>FinOps &amp; Cost Optimisation: Drive fiscal responsibility by right-sizing GCP resources and optimising Generative AI architectures (token management/model selection) to maximise ROI.</li>
</ul>
<ul>
<li>SRE &amp; Performance Tuning: Apply Site Reliability Engineering principles to ensure high availability, minimise system latency, and lead root-cause analysis for complex, distributed system failures.</li>
</ul>
<ul>
<li>DevSecOps &amp; Problem Solving: Integrate security automation into CI/CD pipelines to ensure &#39;Secure-by-Design&#39; deployments while solving complex architectural trade-offs between speed, scale, and risk.</li>
</ul>
<ul>
<li>Continuous Learning: Stay at the forefront of AI research, specifically regarding autonomous agents, prompt engineering etc</li>
</ul>
<p><strong>Nice to Have</strong></p>
<ul>
<li>AI development tools and frameworks (e.g., LangChain, LangGraph, or Agent Dev Kit) to accelerate the delivery of intelligent applications.</li>
</ul>
<ul>
<li>Agentic &amp; GenAI Design: Lead the architectural design of Agentic AI systems (multi-agent orchestration) and Generative AI solutions, including Retrieval-Augmented Generation (RAG) patterns and LLM integration.</li>
</ul>
<ul>
<li>Kubernetes (GKE): Experience managing containerised workloads at scale.</li>
</ul>
<ul>
<li>Kafka/Event-Driven Design: Experience with high-throughput messaging and event-driven architectures.</li>
</ul>
<ul>
<li>MLOps: Familiarity with the end-to-end lifecycle of machine learning models in production.</li>
</ul>
<p><strong>Qualifications</strong></p>
<p><strong>You&#39;ll have...</strong></p>
<ul>
<li>Requires a bachelor&#39;s or foreign equivalent degree in computer science, information technology or a technology related field</li>
</ul>
<ul>
<li>5+ years of Software engineering experience using Java or Python developing services (APIs, REST, etc.)</li>
</ul>
<ul>
<li>2+ years of experience with Google Cloud Platform or other cloud service provider (AWS, Azure, etc.) and associated cloud components.</li>
</ul>
<ul>
<li>Experience designing/architecting and running distributed systems in a production environment</li>
</ul>
<ul>
<li>STRONG communications skills and cognitive agility - ability to engage in deep technical discussions with customers and peers, become a trusted technical advisor, and maintain good documentation</li>
</ul>
<p><strong>Even better, you may have...</strong></p>
<ul>
<li>Master&#39;s degree in computer science, electrical engineering or a closely related field of study</li>
</ul>
<ul>
<li>Familiarity with a breadth of programming languages, platforms, and systems</li>
</ul>
<ul>
<li>Experience with asynchronous messaging and eventually consistent system design</li>
</ul>
<ul>
<li>An agile, pragmatic, and empirical mindset</li>
</ul>
<ul>
<li>Critical thinking, decision-making and leadership aptitudes</li>
</ul>
<ul>
<li>Good organisational and problem-solving abilities</li>
</ul>
<ul>
<li>MDM, Entity Resolution, Customer Analytics and Marketing Analytics experience is a huge plus.</li>
</ul>
<p>You may not check every box, or your experience may look a little different from what we&#39;ve outlined, but if you think you can bring value to Ford Motor Company, we encourage you to apply!</p>
<p><strong>As an established global company, we offer the benefit of choice. You can choose what your Ford future will look like: will your story span the globe, or keep you close to home? Will your career be a deep dive into what you love, or a series of new teams and new skills? Will you be a leader, a changemaker, a technical expert, a culture builder…or all of the above? No matter what you choose, we offer a work life that works for you, including:</strong></p>
<ul>
<li>Immediate medical, dental, and prescription drug coverage</li>
</ul>
<ul>
<li>Flexible family care, parental leave, new parent ramp-up programs, subsidised back-up child care and more</li>
</ul>
<ul>
<li>Vehicle discount programme for employees and family members, and management leases</li>
</ul>
<ul>
<li>Tuition assistance</li>
</ul>
<ul>
<li>Established and active employee resource groups</li>
</ul>
<ul>
<li>Paid time off for individual and team community service</li>
</ul>
<ul>
<li>A generous schedule of paid holidays, including the week between Christmas and New Year&#39;s Day</li>
</ul>
<ul>
<li>Paid time off and the option to purchase additional vacation time.</li>
</ul>
<p><strong>For a detailed look at our benefits, click here:</strong> Benefit Summary</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>senior</Experiencelevel>
      <Workarrangement>remote</Workarrangement>
      <Salaryrange>$115,000-$192,900</Salaryrange>
      <Skills>Google Cloud Platform, BigQuery, Vertex AI, Dataflow, Dataproc, Cloud Run, CloudSQL, Spanner, Apigee, Security &amp; Networking, IAM, Encryption, Complex Network Topologies, Data Management, Enterprise Data Platforms, Data Mesh Architecture, Data-Driven Architectural Patterns, DevOps Tooling, GitHub, SonarQube, Checkmarx, FOSSA, Software Engineering, Web Services, Clean Code Standards, System Design, Horizontal Scalability, Resilience, Clean Code Compliance, Product Collaboration, Agile/Sprint Methodology, Architectural Governance, Cloud Architecture, DevOps Excellence, CI/CD Pipelines, Automated Testing/Security Tools, Secure Infrastructure, Secure-by-Design Principles, Cloud Services, Advanced Networking, VPCs, Service Controls, Load Balancing, Private Service Connect, Cyber Security Oversight, Security Scanning, Compliance, Distributed Systems, AI Models, API &amp; Data Contracts, Data as a Product, API Standards, Data Contracts, Seamless Interoperability, FinOps &amp; Cost Optimisation, Fiscal Responsibility, GCP Resources, Generative AI Architectures, Token Management, Model Selection, ROI Maximisation, SRE &amp; Performance Tuning, High Availability, System Latency, Root-Cause Analysis, DevSecOps &amp; Problem Solving, Security Automation, Continuous Learning, AI Research, Autonomous Agents, Prompt Engineering, Kubernetes, Containerised Workloads, Kafka/Event-Driven Design, High-Throughput Messaging, Event-Driven Architectures, MLOps, Machine Learning Models, End-to-End Lifecycle, AI Development Tools, Frameworks, LangChain, LangGraph, Agent Dev Kit, Agentic &amp; GenAI Design, Multi-Agent Orchestration, Generative AI Solutions, Retrieval-Augmented Generation, LLM Integration, Kubernetes (GKE)</Skills>
      <Category>Engineering</Category>
      <Industry>Automotive</Industry>
      <Employername>Ford Motor Company</Employername>
      <Employerlogo>https://logos.yubhub.co/corporate.ford.com.png</Employerlogo>
      <Employerdescription>Ford Motor Company is a multinational automaker headquartered in Dearborn, Michigan. It is one of the largest automobile manufacturers in the world.</Employerdescription>
      <Employerwebsite>https://corporate.ford.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://efds.fa.em5.oraclecloud.com/hcmUI/CandidateExperience/en/sites/CX_1/job/62370</Applyto>
      <Location>Dearborn</Location>
      <Country></Country>
      <Postedate>2026-04-24</Postedate>
    </job>
    <job>
      <externalid>aebaacf5-640</externalid>
      <Title>Integrations Engineer</Title>
      <Description><![CDATA[<p>You will own the full lifecycle of integrations that power Hebbia&#39;s AI , from designing connectors to deploying them in production, monitoring their behavior, and debugging failures in real time.</p>
<p>You&#39;ll work across systems like Snowflake, S3, SharePoint, and internal customer infrastructure , building pipelines that need to handle real-world complexity: unreliable APIs, evolving schemas, massive datasets, and edge cases that don’t show up in documentation.</p>
<p>This role is hands-on, high-ownership, and deeply technical. You won’t just write code , you’ll develop the instincts to operate and debug complex distributed systems in production.</p>
<p>You will build connectors and ingestion pipelines that bring enterprise data into Hebbia&#39;s AI platform, from Snowflake warehouses and SharePoint libraries to live pricing feeds, high-velocity news data, and proprietary customer systems.</p>
<p>You will design and operate pipelines that handle scale, failures, and edge cases gracefully.</p>
<p>You will debug issues across APIs, auth systems, and data formats, often under real-time customer pressure.</p>
<p>You will own reliability end-to-end: monitoring, alerting, on-call, and incident response.</p>
<p>You will improve internal tooling and observability to make systems more robust and easier to operate.</p>
<p>You will partner with product and customer teams to scope, prioritize, and ship the integrations that unlock Hebbia&#39;s highest-value use cases.</p>
<p>You will design and ship agents that sit on top of the ingestion layer, making enterprise data accessible and actionable across all of Hebbia&#39;s product surfaces , from document analysis to structured query workflows.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>onsite</Workarrangement>
      <Salaryrange>$160,000 to $265,000</Salaryrange>
      <Skills>Python, APIs, OAuth flows, webhook patterns, rate limiting, pagination, cloud infrastructure, AWS, Kafka, PostgreSQL, Redis, ElasticSearch, enterprise data platforms, document processing pipelines, content extraction systems, agentic systems, LLM-enabled products, AI tools</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Hebbia</Employername>
      <Employerlogo>https://logos.yubhub.co/hebbia.com.png</Employerlogo>
      <Employerdescription>Hebbia is an AI platform for investors and bankers that generates alpha and drives upside, founded in 2020 by George Sivulka and backed by Peter Thiel and Andreessen Horowitz.</Employerdescription>
      <Employerwebsite>https://hebbia.com/</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://job-boards.greenhouse.io/hebbia/jobs/4675784005</Applyto>
      <Location>New York City; San Francisco, CA</Location>
      <Country></Country>
      <Postedate>2026-04-17</Postedate>
    </job>
  </jobs>
</source>