{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/title/content-integrity-analyst"},"x-facet":{"type":"title","slug":"content-integrity-analyst","display":"Content Integrity Analyst","count":1},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_0b164900-4da"},"title":"Content Integrity Analyst","description":"<p><strong>Content Integrity Analyst</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Department</strong></p>\n<p>Trust &amp; Safety Operations</p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$252K – $280K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<p><strong>Benefits</strong></p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p><strong>About the Team</strong></p>\n<p>Trust &amp; Safety Operations is central to protecting OpenAI’s platform, customers, and the public from abuse. We support a diverse customer base -- from individual users and early-stage startups to global enterprises -- across ChatGPT, our API, and new product surfaces as they launch.</p>\n<p>Within the Support organization, we partner closely with Product, Engineering, Legal, Policy, Go To Market, and Operations teams to deliver a great user experience at scale while reducing material harm and mitigating catastrophic risks.</p>\n<p><strong>About the Role</strong></p>\n<p>We’re hiring experienced Trust &amp; Safety / Content Integrity operators who can investigate complex cases, apply and evolve usage policy in real-world scenarios, and help build scalable systems that reduce risk over time. You will contribute as a subject-matter expert (SME) on high-stakes escalations, partnering with cross-functional stakeholders to drive fast, defensible outcomes. You will also help design the processes, tooling, and automation that power safe operations at scale.</p>\n<p>This role is ideal for someone who combines strong judgment with sharp analytical instincts, and who enjoys turning ambiguity into clear decisions, repeatable workflows, and durable automation.</p>\n<p><strong>Please note</strong></p>\n<p>This role may involve handling sensitive content, including material that may be highly confidential, sexual, violent, or otherwise disturbing.</p>\n<p><strong>Location</strong></p>\n<p>San Francisco, CA (hybrid: 3 days in office/week)</p>\n<p><strong>In this role you will:</strong></p>\n<ul>\n<li><strong>Apply usage policy with rigor and nuance:</strong> Interpret and apply OpenAI’s usage policies to complex, novel scenarios; provide clear guidance to customers and internal teams; document edge cases and propose policy refinements.</li>\n</ul>\n<ul>\n<li><strong>Mitigate material harm and catastrophic risks:</strong> Triage, assess, and support actions on content and behavior that can drive real-world harm, including high-severity domains; escalate appropriately and help drive cases to resolution.</li>\n</ul>\n<ul>\n<li><strong>Serve as an escalation SME for high-stakes cases:</strong> Support incident response and executive-visible escalations by producing clear assessments, recommending next steps, and coordinating with Legal, Compliance, Security, Product, and Engineering as needed.</li>\n</ul>\n<ul>\n<li><strong>Build scalable trust workflows:</strong> Design and operate processes for human-in-the-loop labeling, content/user reporting, appeals, enforcement actions, and continuous QA -- with a high bar for quality and consistency.</li>\n</ul>\n<ul>\n<li><strong>Drive automation and operational efficiency:</strong> Identify repeatable patterns, translate them into requirements, and partner with Engineering and Data teams to ship tooling and automation (including LLM-enabled automation) that improves speed, accuracy, and coverage.</li>\n</ul>\n<ul>\n<li><strong>Analyze trends and strengthen feedback loops:</strong> Use quantitative and qualitative analysis to surface emerging abuse patterns, measure policy and tooling performance, and feed insights back into detection systems, product mitigations, and policy updates.</li>\n</ul>\n<ul>\n<li><strong>Raise the quality bar:</strong> Define and monitor KPIs, build calibration and QA programs, iterate on reviewer training, and improve guidelines and tooling based on error analysis.</li>\n</ul>\n<ul>\n<li><strong>Enable internal and external teams:</strong> Create playbooks, SOPs, and training that help partner teams understand our enforcement posture, risk thresholds, and operational philosophy.</li>\n</ul>\n<p><strong>You might thrive in this role if you:</strong></p>\n<ul>\n<li><strong>Build for scale:</strong> You’ve taken workflows from zero to one and then scaled them without sacrificing quality.</li>\n</ul>\n<ul>\n<li><strong>Bring deep Trust &amp; Safety experience:</strong> 5+ years in Trust &amp; Safety, integrity, risk, policy enforcement; experience working with vendors is a plus.</li>\n</ul>\n<ul>\n<li><strong>Have strong judgment under ambiguity:</strong> You can make defensible calls in gray areas, write clearly, and adjust quickly as new information arrives.</li>\n</ul>\n<ul>\n<li><strong>Are analytically strong:</strong> You can assess risk, spot trends, and use data to prioritize problems and evaluate solutions.</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_0b164900-4da","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/77b85828-3d4f-4958-b5cf-4787ca04975a","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$252K – $280K","x-skills-required":["Trust & Safety","Content Integrity","Usage Policy","Risk Assessment","Policy Enforcement","Incident Response","Automation","Operational Efficiency","Data Analysis","Quality Assurance"],"x-skills-preferred":["LLM-enabled automation","Cloud-based tools","Data visualization","Machine learning","Natural language processing"],"datePosted":"2026-03-06T18:33:11.240Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Trust & Safety, Content Integrity, Usage Policy, Risk Assessment, Policy Enforcement, Incident Response, Automation, Operational Efficiency, Data Analysis, Quality Assurance, LLM-enabled automation, Cloud-based tools, Data visualization, Machine learning, Natural language processing","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":252000,"maxValue":280000,"unitText":"YEAR"}}}]}