-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
experts based on Phi 3 model with PEFT library #124
Comments
@pclucas14 can you help with this issue? I also have a PR that creates a library compatible with our library out of PEFT adapters, maybe we can merge that and write an example on how to run Arrow using experts trained with TRL and PEFT Thank you |
@AliEdalat would you be able to share with us the name of your PEFT adapters? |
in #86 , we provide a script in you can reload the model with Once you have the model, you can |
Hey @sordonia, I'm @AliEdalat's teammate, and have recently implemented the Arrow algorithm for our project using the PEFT library, achieving promising results. To do this, I modified the forward path in the bnb.py file (since we use QLoRA in our workflow) and made adjustments to methods in the peft_model.py file, along with some other PEFT library files. I thought you might be interested in a dedicated PEFT adapter for the Arrow algorithm, similar to existing ones like Poly and MHR. Let me know if you'd like any help on this! Thanks! |
It'd be great! Do you plan to open a PR into PEFT? I'd be willing to help |
Yes, I'd be glad to contribute! Which branch should I (we) open a PR on? |
Just to clarify my understanding: you are planning to PR your work into the huggingface PEFT library, right? Or into MTTL library? |
Yes! HuggingFace PEFT library. I initially asked about the branch because I thought there might already be a feature request for implementing Arrow in PEFT. Should I go ahead and open a feature request for this, or would you prefer to handle it? (Sorry for any ambiguity!) |
Hi, Super exciting to hear you are getting promising results! Please go ahead and PR into PEFT. I am happy to help if needed, but you probably have a better intuition on how PEFT works than us :) |
Yes please go ahead an open a feature request, we can jump in the PR when needed! |
(we merged our own PEFT support in |
@sordonia @pclucas14 I'll do it for sure! However, I need to first make sure that my code can almost reproduce the result on the zero-shot datasets, like Piqa, as you've reported in the paper. Furthermore, I think I need to refactor the code, so the Arrow can be added as a Tuner Class in the PEFT library, similar to Polytropon and π-Tuning. |
Hello,
Thank you for your interesting research work.
I have 10 experts trained based on the Phi 3 model (datasets selected based on paper cluttering). I have used the TRL and PEFT libraries for training, ensuring the checkpoint structures are suitable for these libraries.
In training the experts, I used LoRA in 4-bit quantized mode. Additionally, I utilized the o and kqv attention in each layer during training.
I would like to know how I can use your code to execute Arrow for merging these experts for each token in every model layer.
I have some errors in the code.
please explain step by step. I am a beginner in this field.
Thank you, and I would appreciate your response.
The text was updated successfully, but these errors were encountered: