This project is mirrored from https://github.com/meta-llama/llama-recipes.
Pull mirroring failed .
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer or owner.
Last successful update .
Repository mirroring has been paused due to too many failed attempts. It can be resumed by a project maintainer or owner.
Last successful update .
- Jan 15, 2025
-
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
Igor Kasianenko authored
-
Igor Kasianenko authored
-
- Jan 10, 2025
-
-
Alexandre Bassel authored
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
- Jan 09, 2025
-
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
Sanyam Bhutani authored
-
- Jan 02, 2025
-
-
Alex Schrimpf authored
-
- Nov 19, 2024
-
-
Guanghui Qin authored
Fix a typo. The FSDP wrapper should wrap the `MllamaCrossAttentionDecoderLayer`, which was missing.
-
JimChienTW authored
-
JimChienTW authored
-
- Nov 16, 2024
-
-
JimChienTW authored
-
- Oct 24, 2024
-
-
celestinoalan authored
**Problem** Currently, we're val_loss.append(best_val_loss) in each epoch. This is misleading because we're appending the corresponding epoch (not best across epochs) quantities in train_loss, train_prep, and val_prep. This is also inconvenient, as one often would like to plot both train and validation losses as a function of the epochs to look for overfitting. **Solution** val_loss.append(eval_epoch_loss)
-
- Oct 21, 2024
-
-
Kai Wu authored
-
- Oct 18, 2024
-
-
Kai Wu authored
-
- Oct 15, 2024
-
-
Matthias Reso authored
-
Patrik Lambert authored
Set numpy seed in finetuning.py to fix it during finetuning (including in custom_dataset.py) and have it set in functions such as Dataset.train_test_split. This avoids having different train/test splits in different ranks, which may cause NCCL collective operation timeout errors.
-
celestinoalan authored
-
- Oct 14, 2024
-
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
- Oct 12, 2024
-
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
- Oct 11, 2024
-
-
Matthias Reso authored
-
Matthias Reso authored
-
Matthias Reso authored
-
- Oct 08, 2024
-
-
Huang Zhihong authored
-
- Oct 02, 2024
-
-
Lucas Ventura authored
Co-authored-by:
Matthias Reso <13337103+mreso@users.noreply.github.com>
-
- Sep 27, 2024
-
-
Ikko Eltociear Ashimine authored
availble -> available
-