What I'm Looking For

After four years building human evaluation systems at YouTube, I'm ready to go deeper on AI alignment and safety. I'm looking for roles where I can work on the problems that matter most.

Areas I'm Interested In

AI Alignment & Safety

Defining how AI systems should behave. Identifying failure modes. Building evaluation frameworks that catch problems before deployment.

Human Feedback Systems

The "HF" in RLHF. Designing rubrics, curating preference data, ensuring the human judgment layer that shapes model behavior is actually good.

Content Quality for ML

Scaling quality evaluation without losing nuance. Cross-cultural considerations. The gap between what we can measure and what we care about.

Trust & Safety for AI Products

Especially conversational AI and products serving vulnerable populations — kids, education, high-stakes domains.

Types of Roles

I'm exploring:

  • Model Behavior / AI Policy roles at labs (Anthropic, OpenAI, Mistral)
  • Safety & Alignment evaluation
  • RLHF / preference data operations and strategy
  • Trust & Safety for AI products
  • AI product roles with a safety or quality focus

What I Bring

  • 4 years designing quality frameworks and curating preference data at scale
  • Direct experience with conversational AI safety evaluation
  • Track record: 50%+ reduction in inappropriate content, 45% increase in quality engagement
  • Cross-cultural expertise (French/German markets, bilingual)
  • Non-traditional background combining humanities judgment with ML systems

Location

Based in New York City, but open to relocating to San Francisco, Paris, or working remotely for the right opportunity.

If this sounds relevant to what you're working on, I'd love to hear from you.

Get in touch