Discussion about this post

User's avatar
Iqbal Singh's avatar

Hey Ben,

A bit new to fine-tuning, so a silly question: In this notebook, are you training on prompt tokens too because we aren't masking them, is this correct? I've seen a few examples where people do mask instructions and only propagate the loss on response tokens. What's your take on which is preferable? Also, if we're not masking prompt tokens, isn't it the same as continued pre-training?

Expand full comment
1 more comment...

No posts