What I'm Looking For
After four years building human evaluation systems at YouTube, I'm ready to go deeper on AI alignment and safety. I'm looking for roles where I can work on the problems that matter most.
Areas I'm Interested In
AI Alignment & Safety
Defining how AI systems should behave. Identifying failure modes. Building evaluation frameworks that catch problems before deployment.
Human Feedback Systems
The "HF" in RLHF. Designing rubrics, curating preference data, ensuring the human judgment layer that shapes model behavior is actually good.
Content Quality for ML
Scaling quality evaluation without losing nuance. Cross-cultural considerations. The gap between what we can measure and what we care about.
Trust & Safety for AI Products
Especially conversational AI and products serving vulnerable populations — kids, education, high-stakes domains.
Types of Roles
I'm exploring:
- •Model Behavior / AI Policy roles at labs (Anthropic, OpenAI, Mistral)
- •Safety & Alignment evaluation
- •RLHF / preference data operations and strategy
- •Trust & Safety for AI products
- •AI product roles with a safety or quality focus
What I Bring
- •4 years designing quality frameworks and curating preference data at scale
- •Direct experience with conversational AI safety evaluation
- •Track record: 50%+ reduction in inappropriate content, 45% increase in quality engagement
- •Cross-cultural expertise (French/German markets, bilingual)
- •Non-traditional background combining humanities judgment with ML systems
Location
Based in New York City, but open to relocating to San Francisco, Paris, or working remotely for the right opportunity.
If this sounds relevant to what you're working on, I'd love to hear from you.
Get in touch