Jokes aside, has anyone tried fine tuning a model on forum conversations?
By which I mean, for chatbots, the data for fine-tuning basically presents it with examples of dialogs. But if you take any given comment on HN and walk its parents, you basically end up with something like that, except there's more than two people involved (usually). So what if you fed that to LLaMA, and then used the same format for the actual chat?
But was it parsed and reformatted specifically in the "chat format" (i.e. the same as inputs later fed to the model when used as a chatbot)? It can make a surprisingly big difference.
By which I mean, for chatbots, the data for fine-tuning basically presents it with examples of dialogs. But if you take any given comment on HN and walk its parents, you basically end up with something like that, except there's more than two people involved (usually). So what if you fed that to LLaMA, and then used the same format for the actual chat?