{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/machine-learning-safety-or-alignment"},"x-facet":{"type":"skill","slug":"machine-learning-safety-or-alignment","display":"Machine Learning Safety Or Alignment","count":1},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_8385c7dc-bad"},"title":"Researcher, Safety & Privacy","description":"<p>We are seeking a Researcher in Privacy-Preserving Safety to help design and build the next generation of privacy-preserving safety systems for frontier AI models. This role sits at the intersection of AI safety, security, and privacy, with a focus on developing auditable, privacy-first mechanisms that enable robust harm detection and mitigation without exposing sensitive user data.</p>\n<p>You will help define and operationalize frameworks for identifying and addressing frontier risks (e.g., bioweapon instructions, malware creation, suicide/self-harm risks, jailbreaks), while ensuring that privacy guarantees remain intact,even under adversarial conditions.</p>\n<p>This role is central to our long-term goal of scaling our automated privacy-preserving safety systems to mitigate potential harms while minimizing human review.</p>\n<p>You’ll work on foundational problems such as privacy-preserving monitoring, algorithmic auditing, secure enclaves, and adversarially robust safety enforcement protocols, helping ensure that safety systems scale without compromising user trust.</p>\n<p>Design and implement privacy-first architectures for detecting and mitigating harmful model behaviors.</p>\n<p>Build frameworks for auditable private identification of high-risk content (jailbreaks, cyber threats, or weaponization instructions).</p>\n<p>Develop strict, auditable mechanisms triggered only by harm signals.</p>\n<p>Drive the development of automated safety systems that preserve privacy at every level.</p>\n<p>You might thrive in this role if you:</p>\n<p>Are a researcher with deep interest in privacy, security, and AI safety, motivated by building systems that are both trustworthy and effective at scale.</p>\n<p>Hold a PhD or equivalent experience in Computer Science, Cryptography, Security, Machine Learning, or related fields</p>\n<p>Have the ability to translate ambiguous problem spaces into formal frameworks and deployable systems</p>\n<p>Demonstrate proficiency in one or more of:</p>\n<p>Privacy-preserving computation (e.g., secure enclaves, MPC, differential privacy)</p>\n<p>Security and adversarial systems</p>\n<p>Machine learning safety or alignment</p>\n<p>Experience designing robust systems under adversarial threat models</p>\n<p>Have experience with AI safety, jailbreak detection, or model alignment</p>\n<p>Are familiar with privacy-preserving machine learning techniques, algorithmic auditing and/or secure system design</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_8385c7dc-bad","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://openai.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/a0feb59d-e66b-4cc7-a685-7f9393d80fb6","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"Full time","x-salary-range":"$295K – $445K","x-skills-required":["Privacy-preserving computation","Security and adversarial systems","Machine learning safety or alignment","Experience designing robust systems under adversarial threat models","AI safety, jailbreak detection, or model alignment"],"x-skills-preferred":["Privacy-preserving machine learning techniques","Algorithmic auditing and/or secure system design"],"datePosted":"2026-04-24T12:22:44.516Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Privacy-preserving computation, Security and adversarial systems, Machine learning safety or alignment, Experience designing robust systems under adversarial threat models, AI safety, jailbreak detection, or model alignment, Privacy-preserving machine learning techniques, Algorithmic auditing and/or secure system design","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":295000,"maxValue":445000,"unitText":"YEAR"}}}]}