Skip to content

[QA] InternLM2微调时数据的max_token #687

Closed Answered by ZwwWayne
MING-ZCH asked this question in Q&A
Discussion options

You must be logged in to vote

仅供参考:出于效率和效果的 trade-off,我们目前用32K长文本进行了训练,然后外推到200K。

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@MING-ZCH
Comment options

Answer selected by ZwwWayne
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
question Further information is requested
2 participants
Converted from issue

This discussion was converted from issue #681 on January 31, 2024 13:41.