请问在lora微调时,如果embedding layer的长度比tokenizer的词表长度稍大,设置“resize_vocab=True”是会使用embedding layer中原来未被使用的部分,还是把embedding layer的长度增大? #4807
Unanswered
CloudyDory
asked this question in
Q&A
Replies: 1 comment
-
在哪里设置resize_vocab=True |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
最近在使用lora微调一个预训练好的模型,使用和预训练时一样的词表。但lora的输出中有如下warning:
我在预训练时使用的词表包含50280个词,但网络的embedding_size是51200,所以embedding layer中还有一小部分剩余空间。如果设置“resize_vocab=True”,是会使用embedding layer中原来未被使用的部分,还是把embedding layer的长度增大?
Beta Was this translation helpful? Give feedback.
All reactions