How can I update a checkpoint's vocabulary?
New vocabulary can be used to continue training from a checkpoint. Existing vocabulary embeddings will be mapped to the new vocabulary, and new vocabulary tokens will be initialized as usual.
Run eole build_vocab
as usual with the new dataset. New vocabulary files will be created.
Training options to perform vocabulary update are:
-update_vocab
: set this option-reset_optim
: set the value to "states"-train_from
: checkpoint path