-
- Downloads
adding flash attention and xformer memory efficient through PT SDPA
Showing
- README.md 9 additions, 1 deletionREADME.md
- configs/training.py 1 addition, 0 deletionsconfigs/training.py
- docs/inference.md 12 additions, 0 deletionsdocs/inference.md
- docs/mutli_gpu.md 8 additions, 1 deletiondocs/mutli_gpu.md
- inference/chat_completion.py 14 additions, 0 deletionsinference/chat_completion.py
- inference/inference.py 13 additions, 0 deletionsinference/inference.py
- llama_finetuning.py 12 additions, 1 deletionllama_finetuning.py
- requirements.txt 1 addition, 1 deletionrequirements.txt
- utils/train_utils.py 17 additions, 6 deletionsutils/train_utils.py
Loading
Please register or sign in to comment