yeah its not fair to call it a finetune because finetune carries connotation of "there wasnt that much extra compute and data added". RLFT has a lot more added to it as Sasha alluded in his talk https://www.youtube.com/watch?v=md8D8eNj5JM - the x axis is log scale, think about that
the framing here is more about "why would you start from random weights when perfectly good starting weights exist" https://www.latent.space/p/fastai
the framing here is more about "why would you start from random weights when perfectly good starting weights exist" https://www.latent.space/p/fastai