diff --git a/src/llama_recipes/finetuning.py b/src/llama_recipes/finetuning.py index 21d8f1c33065a4bfa5c5fe46427f5589f8293076..702a4aa0c5f5faaad140f0995c935cbee935cb93 100644 --- a/src/llama_recipes/finetuning.py +++ b/src/llama_recipes/finetuning.py @@ -76,11 +76,6 @@ def main(**kwargs): model alone would consume 2+TB cpu mem (70 * 4 * 8). This will add some comms overhead and currently requires latest nightly. """ - v = packaging.version.parse(torch.__version__) - verify_latest_nightly = v.is_devrelease and v.dev >= 20230701 - if not verify_latest_nightly: - raise Exception("latest pytorch nightly build is required to run with low_cpu_fsdp config, " - "please install latest nightly.") if rank == 0: model = LlamaForCausalLM.from_pretrained( train_config.model_name,