-
Notifications
You must be signed in to change notification settings - Fork 94
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for LLaMA-2 #23
Comments
I have solved the data problem, but I ran into a new problem. I used wanda to prune LLaMA-2-13B and got a zero score on rouge-2 of CNN/DM, my perplexity of C4 on unstructured pruning is high to 56050.3008. |
Hi, we just updated the repo supporting pruning LLaMA-2 model, see here for the corresponding command. We also provide the results from our own run. |
@Eric-mingjie Thanks! |
@Eric-mingjie Is the performance of ppl related to the environment, I still get poor results on LLaMA-2. |
I think for |
My environment is transformers 4.34.0.dev0, accelerate 0.24.0.dev0 and I get ppl 146760.7188 and now a lot of cuda errors. |
hmm, can you load the |
OK, i will try it to check. |
This is the output of |
@Eric-mingjie I get ppl wikitext_train 5.171178340911865, wikitext_test 4.883730888366699 on Llama-2-13b with no pruning. |
I think it might be helpful to think in terms of why 'wrapped_layers[name].scaler_row' is the all-0 tensor causing the metric to fail, have you run into this? Looks like something's wrong with the hook. |
😭, I finally found the bug, we need to set pretraining_tp to 1, otherwise, the forward will not be executed and the callback will fail. ppl of llama-2-13b (4:8) on wikitext_train 7.27443265914917, wikitext_test 7.004149913787842 |
That's good to know. I was starting to rerun the code on my end. |
Hello @junzhang-zj ,
I changed the code for the c4 data to the following:
Then, I started getting the following error:
I tried downloading with:
After downloading the whole dataset, I need to change the load_dataset function to call the local files. So I did the following:
Now I am getting the following error:
|
@simlaharma Have you tried downloading directly from the huggingface website and then loading it locally? |
I had a similar issue as you did. check this post, it worked for me. |
can we use Wanda for pruning the last linear layer in Llama 2? |
I couldn't reach 'allenai/c4' on the Hub.
The text was updated successfully, but these errors were encountered: