- different methods of finetuning (full retraining, LoRA)
- different base models
- data sets (e.g. Alpaca)
- objectives (creativity, instruction following)
I haven't seen much discussion about people finetuning an LLM on domain-specific data, e.g.
- medical records
- standup comedy jokes
- internal corporate data
So, are any of you fine-tuning your LLMs using such niche data? I'd love to hear about your experiences and motivations!
Even if you're working with proprietary datasets, I'm still interested. After all, knowing what you're doing won't allow us to duplicate it, as we don't have access to the same data.
Part of the problem is that these are both ill-defined problems (I started out being mad that I was getting so many articles about ‘Arsenal’ and after doing the feature engineering to understand why the classifier I was using then (also bag of words) couldn’t learn I hated soccer and loved the NFL I became a soccer fan.) One of these days I am going to try a crisper classification problem, also I want to try fine-tuning a T5.
For today, Yoshinon trains a new model every day and I am using my old classifier because it never screws up.