Skip to Content

AI Research

Introduction Reinforcement Learning from Human or AI Feedback (RLHF, RLAIF) has become the standard recipe for aligning large language models (LLMs). But as we push into the agentic era — where models call…

Get the latest articles in your inbox.

Get the latest articles in your inbox.