-
Notifications
You must be signed in to change notification settings - Fork 256
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]: Reproducing the score of official microsoft/llmlingua-2-xlm-roberta-large-meetingbank #156
Comments
Hi @xvyaward, thanks for your interest and the very detailed description.
Hope these explanations can help you. |
Is there a standard full training script available? We also expect to train a compressor ourselves, including the word_embedding mentioned earlier. |
Yes! We have provided the experiment code for LLMLingua-2 in ./experiments/llmlingua2. The training data for the compressor is also available at HuggingFace. You can run ./experiments/llmlingua2/data_collection/collect_data.sh first, which will get word labels in the original data and filter out bad samples. Then use the train.sh script in ./experiments/llmlingua2/model_training to train the compressor. You may need to modify the training code to include special tokens during training. |
Hi @pzs19, thank you for your kind reply.
I used temperature=0.0 and top_p=1.0 following the paper, and I believe answers are truncated with "\n" during evaluation, by experiments/llmlingua2/evaluation/metrics.py. However, I still can't reproduce the score from the official llmlingua-2-xlm-roberta-large-meetingbank. The score for in-domain meetingbank_qa has especially dropped significantly, from 73.6 to 68.
So which answer is correct to reproduce the score of llmlingua-2-xlm-roberta-large-meetingbank using the official MeetingBank-LLMCompressed dataset? And if possible, can you share example code that handles special tokens during training? Thank you. |
Hi @pzs19, sorry for the misunderstanding. In the last response, I mean if you want to add special tokens during training, you need to modify our training code. In our experiment, special tokens are not added during training. |
Describe the issue
Following the issue 155, I'm trying to reproduce the results of the official llmlingua-2-xlm-roberta-large-meetingbank model using Mistral-7B as black-box llm.
In specific, I tried to fine-tune the XLM-RoBERTa model with the officially provided dataset, using this train.sh.
Here is my detailed process:
Here are the current issues:
I guess this is relevant to the added special tokens in prompt_compressor.py, but train_roberta.py example does nothing about this, so my fine-tuned model has the same word_embedding weight size with the original RoBERTa ([250002, 1024]).
If the process of training the official model is the same as the process provided as an example here, can you please let me know what needs to be changed in the above process?
Thank you for reading.
The text was updated successfully, but these errors were encountered: