{"version":"0.1","company":{"name":"YubHub","url":"https://yubhub.co","jobsUrl":"https://yubhub.co/jobs/skill/safeguards"},"x-facet":{"type":"skill","slug":"safeguards","display":"Safeguards","count":6},"x-feed-size-limit":100,"x-feed-sort":"enriched_at desc","x-feed-notice":"This feed contains at most 100 jobs (the most recently enriched). For the full corpus, use the paginated /stats/by-facet endpoint or /search.","x-generator":"yubhub-xml-generator","x-rights":"Free to redistribute with attribution: \"Data by YubHub (https://yubhub.co)\"","x-schema":"Each entry in `jobs` follows https://schema.org/JobPosting. YubHub-native raw fields carry `x-` prefix.","jobs":[{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_18b6c565-7bb"},"title":"Sr. Software Development Engineer in Test","description":"<p>About Dialpad ---------------- Dialpad is the AI-native business communications platform. We unify calling, messaging, meetings, and contact center on a single platform - powered by AI that understands every conversation in real time.</p>\n<p>More than 70,000 companies around the globe, including WeWork, Asana, NASDAQ, AAA Insurance, COMPASS Realty, Uber, Randstad, and Tractor Supply, rely on Dialpad to build stronger customer connections using real-time, AI-driven insights.</p>\n<p>We’re now leading the shift to Agentic AI: intelligent agents that don’t just analyse conversations but take action by automating workflows, resolving customer issues, and accelerating revenue in real time. Our DAART initiative (Dialpad Agentic AI in Real Time) is redefining what a communications platform can do.</p>\n<p>Visit dialpad.com to learn more.</p>\n<p>Being a Dialer --------------- AI isn’t just a feature; it’s how our teams do their best work every day. We put powerful AI tools in every employee’s hands so they can move faster, think bigger, and achieve more.</p>\n<p>We believe every conversation matters. And we’ve built the platform that turns those conversations into insight and action, for our customers and ourselves.</p>\n<p>We look for people who are intensely curious and hold themselves to a high bar. Our ambition is significant, and achieving it requires a team that operates at the highest level.</p>\n<p>We seek individuals who embody our core traits: Scrappy, Curious, Optimistic, Persistent, and Empathetic.</p>\n<p>Your role -------- As a Sr. SDET in Agentic QA, you will own the test automation and quality frameworks that support Dialpad’s AI Voice Agent services.</p>\n<p>You will develop automated tests for end-to-end product experiences, from frontend UI to backend services to APIs to audio/text interactions.</p>\n<p>You will test orchestration flows, agent configuration experiences, and guardian safeguards to create robust automated coverage for functionality, performance, reliability, UX, and more.</p>\n<p>In this role, you will develop substantial amounts of automated test infrastructure and partner deeply with the development team to make our fast-growing AI platform more testable, more stable, and more delightful for customers.</p>\n<p>This position is based at one of Dialpad’s Canadian offices and reports to a QA Eng Manager in the United States.</p>\n<p>What you’ll do ------------</p>\n<ul>\n<li>Own end-to-end quality for agentic features and workflows, including strategy, development, execution, and release qualification.</li>\n<li>Design and build automation tooling and frameworks for AI/LLM-driven systems, including prompt flows, agent orchestration, and tool integrations.</li>\n<li>Develop and maintain evaluation frameworks (evals) to measure response quality, accuracy, and hallucination rates.</li>\n<li>Drive automation coverage (80%+ for critical AI workflows) using deterministic + probabilistic validation approaches.</li>\n<li>Integrate AI quality checks into CI/CD pipelines with fast feedback cycles (</li>\n<li>Build tooling for LLM observability and debugging, including prompt tracing and response analysis.</li>\n<li>Partner with Applied AI teams on prompt engineering, model selection, and evaluation strategies.</li>\n<li>Design and execute performance and load tests for AI services (latency, throughput, cost efficiency).</li>\n<li>Identify and mitigate risks related to hallucinations, bias, safety, and edge cases.</li>\n<li>Define and track AI quality KPIs (task success rates, precision/recall, latency, etc.).</li>\n<li>Participate in design and architecture reviews to ensure systems are testable, observable, and resilient.</li>\n<li>Mentor engineers and contribute to raising the bar on AI quality engineering practices.</li>\n</ul>\n<p>What you’ll bring --------------</p>\n<ul>\n<li>5+ years of experience in software engineering or SDET roles with an emphasis on software development.</li>\n<li>Strong programming skills in Python (preferred), Java, or JavaScript.</li>\n<li>Experience testing distributed, cloud-native SaaS systems and APIs.</li>\n<li>Demonstrated proficiency in coding with AI agents to accelerate development and improve code quality.</li>\n<li>Hands-on exposure to LLMs or AI/ML systems (e.g., OpenAI, Claude, Gemini, or similar platforms).</li>\n<li>Understanding of non-deterministic systems and probabilistic testing approaches.</li>\n<li>Experience building test frameworks and scalable automation systems.</li>\n<li>Familiarity with AI evaluation techniques (benchmarking, golden datasets, human-in-the-loop validation).</li>\n<li>Experience with CI/CD pipelines (e.g., Jenkins, GitHub Actions).</li>\n<li>Strong collaboration skills with the ability to work across distributed teams and time zones.</li>\n<li>Bachelor’s degree in Computer Science or equivalent practical experience.</li>\n</ul>\n<p>Backend: Python, Go, Google Cloud Platform, Cloud Run / App Engine, Kubernetes, Datastore, Redis, ElasticSearch.</p>\n<ul>\n<li>Frontend: Vue3, React.</li>\n<li>AI Stack: LLM APIs, LiveKit, prompt orchestration frameworks, evaluation tooling.</li>\n</ul>\n<p>For exceptional talent based in British Columbia, Canada the target base salary range for this position is $150,500-$175,250 CAD.</p>\n<p>Why Join Dialpad ---------------</p>\n<ul>\n<li>Work at the center of the AI transformation in business communications.</li>\n<li>Build and ship agentic AI products that are redefining how companies operate.</li>\n<li>Join a team where AI amplifies every employee’s impact.</li>\n<li>Competitive salary, comprehensive benefits, and real opportunities for growth.</li>\n</ul>\n<p>We believe in investing in our people. Dialpad offers competitive benefits and perks, cutting-edge AI tools, and a robust training program that help you reach your full potential.</p>\n<p>We have designed our offices to be inclusive, offering a vibrant environment to cultivate collaboration and connection.</p>\n<p>Our exceptional culture, repeatedly recognized as a Great Place to Work, ensures that every employee feels valued and empowered to contribute to our collective success.</p>\n<p>Don’t meet every single requirement? If you’re excited about this role and possess the fundamental traits, drive, and strong ambition we seek, but your experience doesn’t meet every qualification, we encourage you to apply.</p>\n<p>Dialpad is an equal-opportunity employer. We are dedicated to creating a community of inclusion and an environment free from discrimination or harassment.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_18b6c565-7bb","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Dialpad","sameAs":"https://dialpad.com","logo":"https://logos.yubhub.co/dialpad.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/dialpad/jobs/8475155002","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$150,500-$175,250 CAD","x-skills-required":["Python","Java","JavaScript","Test automation","Quality frameworks","Agentic AI","Voice Agent services","Orchestration flows","Agent configuration experiences","Guardian safeguards","Functional testing","Performance testing","Reliability testing","UX testing","Cloud-native SaaS systems","APIs","LLMs","AI/ML systems","Non-deterministic systems","Probabilistic testing","Test frameworks","Scalable automation systems","CI/CD pipelines","Jenkins","GitHub Actions","Collaboration","Distributed teams","Time zones","Computer Science","Google Cloud Platform","Cloud Run","App Engine","Kubernetes","Datastore","Redis","ElasticSearch","Vue3","React","LLM APIs","LiveKit","Prompt orchestration frameworks","Evaluation tooling"],"x-skills-preferred":[],"datePosted":"2026-04-18T15:49:44.303Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Vancouver, Canada"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Java, JavaScript, Test automation, Quality frameworks, Agentic AI, Voice Agent services, Orchestration flows, Agent configuration experiences, Guardian safeguards, Functional testing, Performance testing, Reliability testing, UX testing, Cloud-native SaaS systems, APIs, LLMs, AI/ML systems, Non-deterministic systems, Probabilistic testing, Test frameworks, Scalable automation systems, CI/CD pipelines, Jenkins, GitHub Actions, Collaboration, Distributed teams, Time zones, Computer Science, Google Cloud Platform, Cloud Run, App Engine, Kubernetes, Datastore, Redis, ElasticSearch, Vue3, React, LLM APIs, LiveKit, Prompt orchestration frameworks, Evaluation tooling","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":150500,"maxValue":175250,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_40c8ba15-3f2"},"title":"Nuclear Material Control and Accountability (MC&A) Specialist","description":"<p>Standard Nuclear is seeking a Nuclear Material Control and Accountability (MC&amp;A) Specialist to ensure secure management, accurate tracking, and regulatory compliance of all nuclear materials.</p>\n<p>This includes implementing systems that prevent unauthorized access, diversion, or misuse of materials, while maintaining full compliance with national and international nuclear safeguards requirements.</p>\n<p>The MC&amp;A Specialist will develop and maintain accounting and inventory systems, support inspections and audits, and collaborate closely with security, engineering, and operations teams to safeguard nuclear materials.</p>\n<p>Responsibilities:</p>\n<ul>\n<li>Maintain accurate, real-time records of all nuclear materials, including acquisition, usage, transfer, and disposal.</li>\n<li>Conduct regular physical inventories and reconcile against accounting systems to ensure accuracy.</li>\n<li>Implement and oversee tamper-indicating devices, seals, and barriers for nuclear material storage and transport.</li>\n</ul>\n<p>Regulatory Compliance &amp; Reporting:</p>\n<ul>\n<li>Prepare and submit required reports to the DOE, NRC, IAEA, and other relevant authorities.</li>\n<li>Ensure compliance with all federal, state, and international nuclear material control regulations.</li>\n<li>Support regulatory inspections, audits, and performance reviews.</li>\n</ul>\n<p>Security &amp; Safeguards:</p>\n<ul>\n<li>Collaborate with security teams to implement deterrence and protective measures against theft, sabotage, or diversion.</li>\n<li>Develop and enforce access protocols and handling procedures for sensitive materials.</li>\n<li>Investigate discrepancies, losses, or potential diversion incidents and implement corrective actions.</li>\n</ul>\n<p>Cross-functional Collaboration:</p>\n<ul>\n<li>Work closely with engineering, manufacturing, and quality assurance teams to integrate MC&amp;A controls into facility operations.</li>\n<li>Communicate compliance status, risks, and corrective actions clearly to leadership and stakeholders.</li>\n</ul>\n<p>Preferred Qualifications:</p>\n<ul>\n<li>Background in Nuclear Safeguards or related technical discipline; 3+ years of relevant experience.</li>\n<li>Experience with DOE, NRC, or IAEA nuclear material safeguards programs.</li>\n<li>Proven ability to manage nuclear material inventory systems and reporting requirements.</li>\n<li>Familiarity with safeguards technologies such as non-destructive assay (NDA) and tamper-indicating devices.</li>\n<li>Strong organizational, analytical, and communication skills, with ability to work proactively and independently.</li>\n</ul>\n<p>Benefits:</p>\n<ul>\n<li>Health, Dental &amp; Vision Insurance</li>\n<li>Health Savings Account</li>\n<li>Disability and Life Insurance</li>\n<li>401K Plan</li>\n<li>Paid Time Off, Holidays</li>\n</ul>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_40c8ba15-3f2","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Standard Nuclear","sameAs":"https://www.standardnuclear.com/","logo":"https://logos.yubhub.co/standardnuclear.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/standardnuclearinc/jobs/5187566008","x-work-arrangement":"onsite","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":null,"x-skills-required":["Nuclear Safeguards","Regulatory Compliance","Inventory Management","Accounting Systems","Tamper-Indicating Devices","Seals and Barriers","DOE","NRC","IAEA","Non-Destructive Assay","Communication Skills"],"x-skills-preferred":["Background in Nuclear Safeguards or related technical discipline","Experience with DOE, NRC, or IAEA nuclear material safeguards programs","Proven ability to manage nuclear material inventory systems and reporting requirements","Familiarity with safeguards technologies such as non-destructive assay (NDA) and tamper-indicating devices","Strong organizational, analytical, and communication skills"],"datePosted":"2026-04-17T12:59:09.947Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"Oak Ridge, TN"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Manufacturing","skills":"Nuclear Safeguards, Regulatory Compliance, Inventory Management, Accounting Systems, Tamper-Indicating Devices, Seals and Barriers, DOE, NRC, IAEA, Non-Destructive Assay, Communication Skills, Background in Nuclear Safeguards or related technical discipline, Experience with DOE, NRC, or IAEA nuclear material safeguards programs, Proven ability to manage nuclear material inventory systems and reporting requirements, Familiarity with safeguards technologies such as non-destructive assay (NDA) and tamper-indicating devices, Strong organizational, analytical, and communication skills"},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_4e0b9271-cdd"},"title":"Research Engineer / Scientist, Alignment Science","description":"<p><strong>About the role:</strong></p>\n<p>You want to build and run elegant and thorough machine learning experiments to help us understand and steer the behavior of powerful AI systems. You care about making AI helpful, honest, and harmless, and are interested in the ways that this could be challenging in the context of human-level capabilities. You could describe yourself as both a scientist and an engineer. As a Research Engineer on Alignment Science, you&#39;ll contribute to exploratory experimental research on AI safety, with a focus on risks from powerful future systems (like those we would designate as ASL-3 or ASL-4 under our Responsible Scaling Policy), often in collaboration with other teams including Interpretability, Fine-Tuning, and the Frontier Red Team.</p>\n<p>Our blog provides an overview of topics that the Alignment Science team is either currently exploring or has previously explored. Our current topics of focus include...</p>\n<ul>\n<li><strong>Scalable Oversight:</strong> Developing techniques to keep highly capable models helpful and honest, even as they surpass human-level intelligence in various domains.</li>\n</ul>\n<ul>\n<li><strong>AI Control:</strong> Creating methods to ensure advanced AI systems remain safe and harmless in unfamiliar or adversarial scenarios.</li>\n</ul>\n<ul>\n<li><strong>Alignment Stress-testing</strong> <strong>:</strong> Creating model organisms of misalignment to improve our empirical understanding of how alignment failures might arise.</li>\n</ul>\n<ul>\n<li><strong>Automated Alignment Research:</strong> Building and aligning a system that can speed up &amp; improve alignment research.</li>\n</ul>\n<ul>\n<li><strong>Alignment Assessments</strong>: Understanding and documenting the highest-stakes and most concerning emerging properties of models through pre-deployment alignment and welfare assessments (see our Claude 4 System Card), misalignment-risk safety cases, and coordination with third-party evaluators.</li>\n</ul>\n<ul>\n<li><strong>Safeguards Research</strong>: Developing robust defenses against adversarial attacks, comprehensive evaluation frameworks for model safety, and automated systems to detect and mitigate potential risks before deployment.</li>\n</ul>\n<ul>\n<li><strong>Model Welfare:</strong> Investigating and addressing potential model welfare, moral status, and related questions. See our program announcement and welfare assessment in the Claude 4 system card for more.</li>\n</ul>\n<p>_Note: For this role, we conduct all interviews in Python and prefer candidates to be based in the Bay Area._</p>\n<p><strong>Representative projects:</strong></p>\n<ul>\n<li>Testing the robustness of our safety techniques by training language models to subvert our safety techniques, and seeing how effective they are at subvertinng our interventions.</li>\n</ul>\n<ul>\n<li>Run multi-agent reinforcement learning experiments to test out techniques like AI Debate.</li>\n</ul>\n<ul>\n<li>Build tooling to efficiently evaluate the effectiveness of novel LLM-generated jailbreaks.</li>\n</ul>\n<ul>\n<li>Write scripts and prompts to efficiently produce evaluation questions to test models’ reasoning abilities in safety-relevant contexts.</li>\n</ul>\n<ul>\n<li>Contribute ideas, figures, and writing to research papers, blog posts, and talks.</li>\n</ul>\n<ul>\n<li>Run experiments that feed into key AI safety efforts at Anthropic, like the design and implementation of our Responsible Scaling Policy.</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have significant software, ML, or research engineering experience</li>\n</ul>\n<ul>\n<li>Have some experience contributing to empirical AI research projects</li>\n</ul>\n<ul>\n<li>Have some familiarity with technical AI safety research</li>\n</ul>\n<ul>\n<li>Prefer fast-moving collaborative projects to extensive solo efforts</li>\n</ul>\n<ul>\n<li>Pick up slack, even if it goes outside your job description</li>\n</ul>\n<ul>\n<li>Care about the impacts of AI</li>\n</ul>\n<p><strong>Strong candidates may also:</strong></p>\n<ul>\n<li>Have experience authoring research papers in machine learning, NLP, or AI safety</li>\n</ul>\n<ul>\n<li>Have experience with LLMs</li>\n</ul>\n<ul>\n<li>Have experience with reinforcement learning</li>\n</ul>\n<ul>\n<li>Have experience with Kubernetes clusters and complex shared codebases</li>\n</ul>\n<p><strong>Candidates need not have:</strong></p>\n<ul>\n<li>100% of the skills needed to perform the job</li>\n</ul>\n<ul>\n<li>Formal certifications or education credentials</li>\n</ul>\n<p>The annual compensation range for this role is listed below.</p>\n<p>For sales roles, the range provided is the role’s On Target Earnings (&quot;OTE&quot;) range, meaning that the range includes both the sales commissions/sales bonuses target and annual base salary for the role.</p>\n<p>Annual Salary:</p>\n<p>$350,000 \\- $500,000USD</p>\n<p><strong><strong>Logistics</strong></strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</p>\n<p><strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</p>\n<p><strong>We encourage you to apply even if you do not believe you meet every single qualification.</strong> Not all strong candidates will meet every single qualification as listed.  Research shows that people who identify as being from underrepresented groups are more prone to experiencing imposter syndrome and doubting the strength of their candidacy, so we urge you not to exclude yourself prematurely and to submit an application if you&#39;re interested in this work. We think AI systems like the ones we&#39;re building have enormous social and ethical implications. We think this makes representation even more important, and we strive to include a range of diverse perspectives on our team.</p>\n<p><strong>Your safety matters to us.</strong> To protect yourself from potential scams, remember that Anthropic recruits through our website and other job boards, and we will never ask you to pay for any part of the recruitment process.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_4e0b9271-cdd","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/4631822008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$350,000 - $500,000USD","x-skills-required":["Python","Machine Learning","Research Engineering","AI Safety","Scalable Oversight","AI Control","Alignment Stress-testing","Automated Alignment Research","Alignment Assessments","Safeguards Research","Model Welfare"],"x-skills-preferred":["Experience authoring research papers in machine learning, NLP, or AI safety","Experience with LLMs","Experience with reinforcement learning","Experience with Kubernetes clusters and complex shared codebases"],"datePosted":"2026-03-08T13:51:34.613Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"Python, Machine Learning, Research Engineering, AI Safety, Scalable Oversight, AI Control, Alignment Stress-testing, Automated Alignment Research, Alignment Assessments, Safeguards Research, Model Welfare, Experience authoring research papers in machine learning, NLP, or AI safety, Experience with LLMs, Experience with reinforcement learning, Experience with Kubernetes clusters and complex shared codebases","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":350000,"maxValue":500000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_1eb84736-326"},"title":"Research Product Manager, Model Behaviors","description":"<p><strong>About the role:</strong></p>\n<p>As a Product Manager for Model Behaviors, you will partner with the Alignment Finetuning team to define and shape Claude&#39;s character, behaviours, and reinforcement signals—work that directly influences how millions of people experience AI. You will systematically identify high-priority behavioural improvements, coordinate across Research, Product, and Safeguards teams, and accelerate our ability to ship well-aligned models.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Define behavioural defaults and steerability constraints</li>\n<li>Develop and maintain taxonomies of model behaviours across capabilities</li>\n<li>Identify, triage, and prioritise behaviour issues and opportunities, coordinating input from Users, Research, Product, and Safeguards teams</li>\n<li>Amplify alignment research breakthroughs, translating them into product, process, and model improvements</li>\n<li>Deeply understand user interaction patterns to identify behaviour improvements that make Claude more helpful and safe</li>\n<li>Contribute to evals that measure alignment progress</li>\n<li>Identify and scale initiatives and tools that help researchers ship alignment improvements faster</li>\n</ul>\n<p><strong>You might be a good fit if you:</strong></p>\n<ul>\n<li>Have a deep passion and curiosity for AI and LLMs. Use AI regularly.</li>\n<li>Have 5+ years in product management leading scaled conversational AI products.</li>\n<li>Are a first-principles thinker with the ability to navigate and execute amidst ambiguity, flexing into different domains based on the business problem at hand and finding simple, easy-to-understand solutions</li>\n<li>Have a track record of delivering products and features to end-users (consumer or end-user b2b focus)</li>\n<li>Have strong user empathy and the ability to synthesise vague or contradictory feedback into actionable priorities</li>\n<li>Have strong judgment and model taste, with the ability to make tradeoffs when there is no clear right answer</li>\n<li>Have a strong grasp of ML concepts and are willing to go deep on technical solutions</li>\n<li>Have intellectual curiosity without ego—comfortable asking questions and learning independently</li>\n<li>Think creatively about the risks and benefits of new technologies, moving beyond past checklists and playbooks</li>\n<li>Have a creative, hacker spirit and love solving puzzles</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<ul>\n<li>Education requirements: We require at least a Bachelor&#39;s degree in a related field or equivalent experience.</li>\n<li>Location-based hybrid policy: Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices.</li>\n<li>Visa sponsorship: We do sponsor visas! However, we aren&#39;t able to successfully sponsor visas for every role and every candidate. But if we make you an offer, we will make every reasonable effort to get you a visa, and we retain an immigration lawyer to help with this.</li>\n</ul>\n<p><strong>How we&#39;re different</strong></p>\n<p>We believe that the highest-impact AI research will be big science. At Anthropic we work as a single cohesive team on just a few large-scale research efforts. And we value impact — advancing our long-term goals of steerable, trustworthy AI — rather than work on smaller and more specific puzzles. We view AI research as an empirical science, which has as much in common with physics and biology as with traditional efforts in computer science. We&#39;re an extremely collaborative group, and we host frequent research discussions to ensure that we are pursuing the highest-impact work at any given time. As such, we greatly value communication skills.</p>\n<p>The easiest way to understand our research directions is to read our recent research. This research continues many of the directions our team worked on prior to Anthropic, including: GPT-3, Circuit-Based Interpretability, Multimodal Neurons, Scaling Laws, AI &amp; Compute, Concrete Problems in AI Safety, and Learning from Human Preference</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_1eb84736-326","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://www.anthropic.com","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5097067008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$305,000 - $385,000 USD","x-skills-required":["product management","conversational AI","machine learning","research","safeguards","user experience","data analysis","communication"],"x-skills-preferred":["first-principles thinking","problem-solving","collaboration","leadership","project management","technical writing"],"datePosted":"2026-03-08T13:47:46.858Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA | New York City, NY"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"product management, conversational AI, machine learning, research, safeguards, user experience, data analysis, communication, first-principles thinking, problem-solving, collaboration, leadership, project management, technical writing","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":305000,"maxValue":385000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_c33b2d78-cc9"},"title":"Research Lead, Training Insights","description":"<p><strong>About the role</strong></p>\n<p>As a Research Lead on the Training Insights team, you&#39;ll develop the strategy for, and lead execution on, how we measure and characterise model capabilities across training and deployment. This is a hands-on leadership role: you&#39;ll drive original research into new evaluation methodologies while leading a small team of researchers and research engineers doing the same.</p>\n<p>Your work will span the full lifecycle of model development. You&#39;ll research and build new long-horizon evaluations that test the boundaries of what our models can achieve, develop novel approaches to measuring emerging capabilities, and deepen our understanding of how those capabilities develop — both during production RL training and after. You&#39;ll also take a cross-organisational view, working across Reinforcement Learning, Pretraining, Inference, Product, Alignment, Safeguards, and other teams to map the landscape of model evaluations at Anthropic and identify critical gaps in coverage.</p>\n<p>This role carries significant visibility and impact. You&#39;ll help shape the evaluation narrative for model releases, contributing directly to how Anthropic communicates about its models to both internal and external audiences. Done well, you will change how the industry measures and understands model capabilities, significantly furthering our safety mission.</p>\n<p><strong>Responsibilities:</strong></p>\n<ul>\n<li>Build new novel and long-horizon evaluations</li>\n<li>Develop novel measurement approaches for understanding how model capabilities emerge and evolve during RL training</li>\n<li>Lead strategic evaluation coverage across the company</li>\n<li>Shape the evaluation narrative for model releases</li>\n<li>Lead and mentor a small team of researchers and research engineers, setting research direction and fostering a culture of rigorous, creative research</li>\n<li>Design evaluation frameworks that balance scientific rigor with the practical demands of production training schedules</li>\n<li>Build and maintain relationships across Anthropic&#39;s research organisation to ensure evaluation insights inform training and deployment decisions</li>\n<li>Contribute to the broader research community through publications, open-source contributions, or external engagement on evaluation best practices</li>\n</ul>\n<p><strong>You may be a good fit if you:</strong></p>\n<ul>\n<li>Have significant experience designing and running evaluations for large language models or similar complex ML systems</li>\n<li>Have led technical projects or teams, either formally or through sustained ownership of critical research directions</li>\n<li>Are equally comfortable designing experiments and writing code—you can move between research and implementation fluidly</li>\n<li>Think strategically about what to measure and why, not just how to measure it</li>\n<li>Can synthesise information across multiple teams and workstreams to form a coherent picture of model capabilities</li>\n<li>Communicate complex technical findings clearly to both technical and non-technical audiences</li>\n<li>Are results-oriented and thrive in fast-paced environments where priorities shift based on research findings</li>\n<li>Care deeply about AI safety and want your work to directly influence how capable AI systems are developed and deployed</li>\n</ul>\n<p><strong>Strong candidates may also have:</strong></p>\n<ul>\n<li>Experience building evaluations for long-horizon or agentic tasks</li>\n<li>Deep familiarity with Reinforcement Learning training dynamics and how model behaviour changes during training</li>\n<li>Published research in machine learning evaluation, benchmarking, or related areas</li>\n<li>Experience with safety evaluation frameworks and red teaming methodologies</li>\n<li>Background in psychometrics, experimental psychology, or other measurement-focused disciplines</li>\n<li>A track record of communicating evaluation results to inform high-stakes decisions about model development or deployment</li>\n<li>Experience managing or mentoring researchers and engineers</li>\n</ul>\n<p><strong>Representative projects:</strong></p>\n<ul>\n<li>Designing and implementing a suite of long-horizon evaluations that test model capabilities on tasks requiring sustained reasoning, planning, and tool use over extended interactions</li>\n<li>Building systems to track capability development across RL training checkpoints, surfacing insights about when and how specific capabilities emerge</li>\n<li>Conducting a cross-org audit of evaluation coverage, identifying blind spots, and prioritising new evaluations to fill critical gaps across Pretraining, RL, Inference, and Product</li>\n<li>Developing the evaluation methodology and narrative for a major model release, working with research leads and communications to clearly characterise model capabilities and limitations</li>\n<li>Researching and prototyping novel evaluation approaches for capabilities that are difficult to measure with existing benchmarks</li>\n<li>Leading a team effort to build reusable evaluation infrastructure that serves multiple teams across the research organisation</li>\n</ul>\n<p><strong>Logistics</strong></p>\n<p><strong>Education requirements:</strong> We require at least a Bachelor&#39;s degree in a related field or equivalent experience. <strong>Location-based hybrid policy:</strong> Currently, we expect all staff to be in one of our offices at least 25% of the time. However, some roles may require more time in our offices repsectively.</p>\n<p><strong>Visa sponsorship:</strong> We do sponsor visas!</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_c33b2d78-cc9","directApply":true,"hiringOrganization":{"@type":"Organization","name":"Anthropic","sameAs":"https://job-boards.greenhouse.io","logo":"https://logos.yubhub.co/anthropic.com.png"},"x-apply-url":"https://job-boards.greenhouse.io/anthropic/jobs/5139654008","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$850,000 - $850,000USD","x-skills-required":["machine learning","evaluation methodologies","Reinforcement Learning","Pretraining","Inference","Product","Alignment","Safeguards"],"x-skills-preferred":["psychometrics","experimental psychology","safety evaluation frameworks","red teaming methodologies"],"datePosted":"2026-03-08T13:45:37.187Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco, CA"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"machine learning, evaluation methodologies, Reinforcement Learning, Pretraining, Inference, Product, Alignment, Safeguards, psychometrics, experimental psychology, safety evaluation frameworks, red teaming methodologies","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":850000,"maxValue":850000,"unitText":"YEAR"}}},{"@context":"https://schema.org","@type":"JobPosting","identifier":{"@type":"PropertyValue","name":"YubHub","value":"job_2aaccebf-892"},"title":"Fraud & Risk Analyst","description":"<p><strong>Job Posting</strong></p>\n<p><strong>Fraud &amp; Risk Analyst</strong></p>\n<p><strong>Location</strong></p>\n<p>San Francisco</p>\n<p><strong>Employment Type</strong></p>\n<p>Full time</p>\n<p><strong>Department</strong></p>\n<p><strong>Compensation</strong></p>\n<ul>\n<li>$252K – $280K • Offers Equity</li>\n</ul>\n<p>The base pay offered may vary depending on multiple individualized factors, including market location, job-related knowledge, skills, and experience. If the role is non-exempt, overtime pay will be provided consistent with applicable laws. In addition to the salary range listed above, total compensation also includes generous equity, performance-related bonus(es) for eligible employees, and the following benefits.</p>\n<ul>\n<li>Medical, dental, and vision insurance for you and your family, with employer contributions to Health Savings Accounts</li>\n</ul>\n<ul>\n<li>Pre-tax accounts for Health FSA, Dependent Care FSA, and commuter expenses (parking and transit)</li>\n</ul>\n<ul>\n<li>401(k) retirement plan with employer match</li>\n</ul>\n<ul>\n<li>Paid parental leave (up to 24 weeks for birth parents and 20 weeks for non-birthing parents), plus paid medical and caregiver leave (up to 8 weeks)</li>\n</ul>\n<ul>\n<li>Paid time off: flexible PTO for exempt employees and up to 15 days annually for non-exempt employees</li>\n</ul>\n<ul>\n<li>13+ paid company holidays, and multiple paid coordinated company office closures throughout the year for focus and recharge, plus paid sick or safe time (1 hour per 30 hours worked, or more, as required by applicable state or local law)</li>\n</ul>\n<ul>\n<li>Mental health and wellness support</li>\n</ul>\n<ul>\n<li>Employer-paid basic life and disability coverage</li>\n</ul>\n<ul>\n<li>Annual learning and development stipend to fuel your professional growth</li>\n</ul>\n<ul>\n<li>Daily meals in our offices, and meal delivery credits as eligible</li>\n</ul>\n<ul>\n<li>Relocation support for eligible employees</li>\n</ul>\n<ul>\n<li>Additional taxable fringe benefits, such as charitable donation matching and wellness stipends, may also be provided.</li>\n</ul>\n<p>More details about our benefits are available to candidates during the hiring process.</p>\n<p>This role is at-will and OpenAI reserves the right to modify base pay and other compensation components at any time based on individual performance, team or company results, or market conditions.</p>\n<p><strong>About the Team</strong></p>\n<p>The Account &amp; Platform Integrity team protects OpenAI’s ecosystem from fraud, impersonation, abuse, and account-level threats. We ensure that the people and organizations using OpenAI are who they claim to be, that access is used appropriately, and that bad actors are prevented from exploiting the platform.</p>\n<p>We operate at the intersection of identity, access, compliance, and abuse prevention, working closely with Product, Engineering, Legal, Go-To-Market, and Support teams to stop harmful activity before it impacts users, customers, or the business. Our work directly protects revenue, user trust, and platform safety across ChatGPT, the API, and enterprise products.</p>\n<p><strong>About the Role</strong></p>\n<p>We’re hiring a Fraud &amp; Risk Analyst to help safeguard OpenAI by investigating, validating, and monitoring customer accounts and organizations. You will focus on identity, legitimacy, and risk, ensuring accounts are properly verified, access is appropriate, and emerging threats are detected early.</p>\n<p>You’ll handle sensitive and high-stakes investigations involving fraud, impersonation, sanctions, misuse of access, and coordinated abuse. Your work will directly influence who can use OpenAI’s products and how safely we can scale.</p>\n<p>_<strong>Note: This role may involve reviewing sensitive, confidential, or disturbing content.</strong>_</p>\n<p>We use a hybrid work model of 3 days in the office per week in our San Francisco office.</p>\n<p><strong>In this role you will:</strong></p>\n<ul>\n<li>Review and verify customer identities, organizations, and ownership structures</li>\n</ul>\n<ul>\n<li>Investigate suspicious or high-risk accounts (e.g., fraud, impersonation, shell companies, abuse of API or ChatGPT access)</li>\n</ul>\n<ul>\n<li>Evaluate documents, internal data, and third-party sources to determine legitimacy and risk</li>\n</ul>\n<ul>\n<li>Enforce account-level actions such as approvals, restrictions, suspensions, or escalations</li>\n</ul>\n<ul>\n<li>Serve as the case owner for complex, high-visibility verification and integrity cases</li>\n</ul>\n<ul>\n<li>Partner with Legal, Compliance, Sales, and Support to resolve issues quickly and accurately</li>\n</ul>\n<ul>\n<li>Handle escalations, appeals, and sensitive customer communications</li>\n</ul>\n<ul>\n<li>Help design and improve verification workflows, fraud detection, and risk-scoring systems</li>\n</ul>\n<ul>\n<li>Contribute to automation, tooling, and human-in-the-loop review pipelines</li>\n</ul>\n<ul>\n<li>Identify patterns of abuse and recommend new controls or safeguards</li>\n</ul>\n<ul>\n<li>Analyze data to uncover fraud and abuse trends</li>\n</ul>\n<ul>\n<li>Provide feedback to Product and Engineering to improve onboarding, verification, and access controls</li>\n</ul>\n<ul>\n<li>Create clear playbooks and guidance for frontline teams handling high-risk accounts</li>\n</ul>\n<p><strong>You Might Thrive In This Role If You...</strong></p>\n<ul>\n<li>Have 5+ years of experience in verifications, fraud, trust &amp; safety, or risk investigations</li>\n</ul>\n<ul>\n<li>Are comfortable making high-impact decisions about who should — or should not — have platform access</li>\n</ul>\n<ul>\n<li>Have experience working cross-functionally with Legal, Product, Sales, and Operations</li>\n</ul>\n<ul>\n<li>Enjoy building systems, not just running them — especially in fast-moving environments</li>\n</ul>\n<ul>\n<li>Are calm under pressure, detail-oriented, and trusted with sensitive and ambiguous cases</li>\n</ul>\n<ul>\n<li>Thrive in environments that require judgment, speed, and accountability</li>\n</ul>\n<p><strong>About OpenAI</strong></p>\n<p>OpenAI is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity. We push the boundaries of the capabilities of AI systems and seek to safely deploy them to the world through our products. AI is an extremely powerful tool that must be created with safety and human needs at its core, and to achieve our mission, we must encompass and value the many different perspectives, voices, and experiences that form the full spectrum of humanity.</p>\n<p style=\"margin-top:24px;font-size:13px;color:#666;\">XML job scraping automation by <a href=\"https://yubhub.co\">YubHub</a></p>","url":"https://yubhub.co/jobs/job_2aaccebf-892","directApply":true,"hiringOrganization":{"@type":"Organization","name":"OpenAI","sameAs":"https://jobs.ashbyhq.com","logo":"https://logos.yubhub.co/openai.com.png"},"x-apply-url":"https://jobs.ashbyhq.com/openai/2f13a80f-645e-44a6-9af8-f183d3409203","x-work-arrangement":"hybrid","x-experience-level":"senior","x-job-type":"full-time","x-salary-range":"$252K – $280K • Offers Equity","x-skills-required":["fraud","risk","investigations","identity","legitimacy","risk","access","compliance","abuse","prevention","data","analysis","trends","fraud","detection","risk-scoring","systems","automation","tooling","human-in-the-loop","review","pipelines","patterns","abuse","controls","safeguards","onboarding","verification","access","controls"],"x-skills-preferred":[],"datePosted":"2026-03-06T18:33:12.581Z","jobLocation":{"@type":"Place","address":{"@type":"PostalAddress","addressLocality":"San Francisco"}},"employmentType":"FULL_TIME","occupationalCategory":"Engineering","industry":"Technology","skills":"fraud, risk, investigations, identity, legitimacy, risk, access, compliance, abuse, prevention, data, analysis, trends, fraud, detection, risk-scoring, systems, automation, tooling, human-in-the-loop, review, pipelines, patterns, abuse, controls, safeguards, onboarding, verification, access, controls","baseSalary":{"@type":"MonetaryAmount","currency":"USD","value":{"@type":"QuantitativeValue","minValue":252000,"maxValue":280000,"unitText":"YEAR"}}}]}