2 Apr
2022
2 Apr
'22
3:06 p.m.
detokenization-training time. let's make a finetuning loop. since the model is just like e.g. gpt-2, it should be reaosnable to train. set the parts to train, load an optimizer, calculate a loss, backpropagate the gradients. we'll need input data. not sure what to use. maybe i'll feed it random noise or something at first. two models. one runs with original detokenizer, another runs with new one. adjust new model to match old output. lots of parts. >5 lines. are copyable if more pages are loaded. On 4/2/22, Undiscussed Horrific Abuse, One Victim of Many <gmkarl@gmail.com> wrote:
[we are used to doing other behavior. sad that we are living this pattern while it is being reduced.]