Original Reddit post

It seems pretty well established that Claude is heads above its immediate competition. Was wondering two things:

  • Why?
  • Where the training data actually comes from? I would think the bulk of code trainable would be directly from Github. A very basic high-level process would probably be Github code -> base model -> RLHF for the instruct model. Sensible opinion would be ‘maybe Claude has stronger RLHF processes’ or something. But I am wondering if Anthropic actually does use different base corpora from other models. Is anyone more savvy than me able to comment on this? submitted by /u/MullingMulianto

Originally posted by u/MullingMulianto on r/ClaudeCode