Anthropic is making some big changes to how it handles user data, requiring all Claude users to decide by September 28 whether they want their conversations used to train AI models. While the company directed us to its blog post on the policy changes when asked about what prompted the move, we've formed some theories of our own. But first, what's changing: previously, Anthropic didn't use consumer chat data for model training.
In our head-to-head experiments, we observe that offline contrastive training offers a more valuable training signal than traditional SFT methods, demonstrating its effectiveness in model performance.