-
- Downloads
adding flash attention and xformer memory efficient through PT SDPA (#97)
Showing
- README.md 9 additions, 1 deletionREADME.md
- configs/training.py 1 addition, 0 deletionsconfigs/training.py
- docs/inference.md 12 additions, 0 deletionsdocs/inference.md
- docs/multi_gpu.md 8 additions, 1 deletiondocs/multi_gpu.md
- inference/chat_completion.py 13 additions, 0 deletionsinference/chat_completion.py
- inference/inference.py 18 additions, 5 deletionsinference/inference.py
- llama_finetuning.py 11 additions, 2 deletionsllama_finetuning.py
- requirements.txt 1 addition, 1 deletionrequirements.txt
- scripts/spellcheck_conf/wordlist.txt 30 additions, 1 deletionscripts/spellcheck_conf/wordlist.txt
- utils/train_utils.py 17 additions, 6 deletionsutils/train_utils.py
Loading
Please register or sign in to comment