FSDP+LORA on multiple gpu(A100 80gb*4) ValueError: Cannot flatten integer dtype tensors #2250
Open
6 of 8 tasks
Labels
bug
Something isn't working
Please check that this issue hasn't been reported before.
Expected Behavior
The LoRA configuration should work with fsdp
Current behaviour
[rank0]: raise ValueError("Cannot flatten integer dtype tensors")
[rank0]: ValueError: Cannot flatten integer dtype tensors
[rank1]: Traceback (most recent call last):
Steps to reproduce
Run Axolotl on multiple GPUs using LoRA+ FSDP, 4 NVIDIA A100 GPUs with 80GB
-torch Version: 2.5.1
-axolotl Version: 0.6.0
Config yaml
Possible solution
No response
Which Operating Systems are you using?
Python Version
3.10.13
axolotl branch-commit
main/latest
Acknowledgements
The text was updated successfully, but these errors were encountered: