<?xml version="1.0" encoding="UTF-8"?>
<source>
  <jobs>
    <job>
      <externalid>58b03260-1e2</externalid>
      <Title>AI Engineer, Product</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>
<p>We are a global company with a diverse workforceREADME</p>
<p>Embedded directly in a product team as search, chat, documents, or audio, you&#39;ll improve AI-powered features through rigorous evaluation, prompt and orchestration design, and rapid experimentation. You&#39;ll own your domain&#39;s AI quality end-to-end: define what &quot;good&quot; looks like, measure it, run experiments, and ship what works.</p>
<p>Responsibilities</p>
<p>• Design and run evaluations for your product area: reference tests, heuristics, model-graded checks tailored to search relevance, chat quality, document understanding, or audio performance.</p>
<p>• Define and track metrics that matter: task success, helpfulness, hallucination proxies, safety flags, latency, cost.</p>
<p>• Own prompt and orchestration design: write, test, and iterate on prompts and system prompts as a core part of your work.</p>
<p>• Run A/B tests on prompts, models, and configurations; analyze results; make rollout or rollback decisions from data.</p>
<p>• Set up observability for LLM calls: structured logging, tracing, dashboards, alerts.</p>
<p>• Operate model releases: canary and shadow traffic, sign-offs, SLO-based rollback criteria, regression detection.</p>
<p>• Improve core behaviors in your product area, whether that&#39;s memory policies, intent classification, routing, tool-call reliability, or retrieval quality.</p>
<p>• Create templates and documentation so other teams can author evals and ship safely.</p>
<p>• Partner with Science to diagnose regressions and lead post-mortems.</p>
<p>About you</p>
<p>• 3-4 years of experience; backgrounds that fit well include ML engineers moving closer to product, or software engineers with real AI/ML production experience.</p>
<p>• Strong TypeScript or Python skills - we have both tracks depending on team fit.</p>
<p>• Production LLM experience: prompts, tool/function calling, system prompts.</p>
<p>• Hands-on with evals and A/B testing; you can design metrics, not just run them.</p>
<p>• Comfortable implementing directly in product code, not only notebooks.</p>
<p>• Observability experience: logging, tracing, dashboards, alerting.</p>
<p>• Product mindset: form hypotheses, run experiments, interpret results, ship.</p>
<p>• Clear communication, autonomous, and oriented toward production impact over experimentation for its own sake.</p>
<p>It would be ideal if you also have:</p>
<p>• Safety systems experience: moderation, PII handling/redaction, guardrails.</p>
<p>• Release operations: canary/shadowing, automated rollbacks, experiment platforms.</p>
<p>• Prior work on search ranking, chat systems, document AI, or audio ML features.</p>
<p>Hiring Process</p>
<p>• Introduction call - 30 min</p>
<p>• Hiring Manager interview - 30 min</p>
<p>• Technical Rounds - Live-coding Interview - 45 min - AI Engineering Interview - 45 min</p>
<p>• Culture-fit discussion - 30 min</p>
<p>• References</p>
<p>By applying, you agree to our Applicant Privacy Policy.</p>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>TypeScript, Python, Production LLM experience, Evals and A/B testing, Observability, Product mindset, Clear communication, Safety systems experience, Release operations, Search ranking, Chat systems, Document AI, Audio ML features</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo>https://logos.yubhub.co/mistral.ai.png</Employerlogo>
      <Employerdescription>Mistral AI develops high-performance, open-source AI models and solutions for enterprise use. Its comprehensive AI platform meets needs in both on-premises and cloud environments.</Employerdescription>
      <Employerwebsite>https://mistral.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/c79ff8ed-6689-4dda-aec6-979a5dc767d0</Applyto>
      <Location>Paris</Location>
      <Country></Country>
      <Postedate>2026-04-17</Postedate>
    </job>
    <job>
      <externalid>6663d8f4-ea5</externalid>
      <Title>AI Engineer, Product</Title>
      <Description><![CDATA[<p>About Mistral AI</p>
<p>At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.</p>
<p>We are a global company with teams distributed between France, USA, UK, Germany, and Singapore. Our diverse workforce thrives in competitive environments and is committed to driving innovation.</p>
<p>Role Summary</p>
<p>Embedded directly in a product team as search, chat, documents, or audio, you&#39;ll improve AI-powered features through rigorous evaluation, prompt and orchestration design, and rapid experimentation. You&#39;ll own your domain&#39;s AI quality end-to-end: define what &#39;good&#39; looks like, measure it, run experiments, and ship what works.</p>
<p>Responsibilities</p>
<ul>
<li>Design and run evaluations for your product area: reference tests, heuristics, model-graded checks tailored to search relevance, chat quality, document understanding, or audio performance.</li>
<li>Define and track metrics that matter: task success, helpfulness, hallucination proxies, safety flags, latency, cost.</li>
<li>Own prompt and orchestration design: write, test, and iterate on prompts and system prompts as a core part of your work.</li>
<li>Run A/B tests on prompts, models, and configurations; analyze results; make rollout or rollback decisions from data.</li>
<li>Set up observability for LLM calls: structured logging, tracing, dashboards, alerts.</li>
<li>Operate model releases: canary and shadow traffic, sign-offs, SLO-based rollback criteria, regression detection.</li>
<li>Improve core behaviors in your product area, whether that&#39;s memory policies, intent classification, routing, tool-call reliability, or retrieval quality.</li>
<li>Create templates and documentation so other teams can author evals and ship safely.</li>
<li>Partner with Science to diagnose regressions and lead post-mortems.</li>
</ul>
<p>About You</p>
<ul>
<li>3-4 years of experience; backgrounds that fit well include ML engineers moving closer to product, or software engineers with real AI/ML production experience.</li>
<li>Strong TypeScript or Python skills - we have both tracks depending on team fit.</li>
<li>Production LLM experience: prompts, tool/function calling, system prompts.</li>
<li>Hands-on with evals and A/B testing; you can design metrics, not just run them.</li>
<li>Comfortable implementing directly in product code, not only notebooks.</li>
<li>Observability experience: logging, tracing, dashboards, alerting.</li>
<li>Product mindset: form hypotheses, run experiments, interpret results, ship.</li>
<li>Clear communication, autonomous, and oriented toward production impact over experimentation for its own sake.</li>
</ul>
<p>Benefits</p>
<ul>
<li>Competitive salary and equity package</li>
<li>Health insurance</li>
<li>Transportation allowance</li>
<li>Sport allowance</li>
<li>Meal vouchers</li>
<li>Private pension plan</li>
<li>Generous parental leave policy</li>
</ul>
<p style="margin-top:24px;font-size:13px;color:#666;">XML job scraping automation by <a href="https://yubhub.co">YubHub</a></p>]]></Description>
      <Jobtype>full-time</Jobtype>
      <Experiencelevel>mid</Experiencelevel>
      <Workarrangement>hybrid</Workarrangement>
      <Salaryrange></Salaryrange>
      <Skills>TypeScript, Python, Production LLM experience, Evals and A/B testing, Observability, Product mindset, Safety systems experience, Release operations, Search ranking, Chat systems, Document AI, Audio ML features</Skills>
      <Category>Engineering</Category>
      <Industry>Technology</Industry>
      <Employername>Mistral AI</Employername>
      <Employerlogo></Employerlogo>
      <Employerdescription>Mistral AI develops high-performance, open-source AI models and solutions for enterprise use.</Employerdescription>
      <Employerwebsite>https://mistral.ai</Employerwebsite>
      <Compensationcurrency></Compensationcurrency>
      <Compensationmin></Compensationmin>
      <Compensationmax></Compensationmax>
      <Applyto>https://jobs.lever.co/mistral/c79ff8ed-6689-4dda-aec6-979a5dc767d0</Applyto>
      <Location>Paris</Location>
      <Country></Country>
      <Postedate>2026-03-10</Postedate>
    </job>
  </jobs>
</source>