Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

remove default to ignore safetensors #2042

Merged
merged 12 commits into from
Nov 21, 2024

Conversation

felipemello1
Copy link
Contributor

Context

What is the purpose of this PR? Is it to

  • add a new feature
  • fix a bug
  • update tests and/or documentation
  • other (please add here)

bad default

Testing:
Downloaded models

Copy link

pytorch-bot bot commented Nov 21, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchtune/2042

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

✅ No Failures

As of commit b05689b with merge base e9fd56a (image):
💚 Looks good so far! There are no failures yet. 💚

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Nov 21, 2024
Felipe Mello added 2 commits November 21, 2024 08:53
@@ -99,7 +99,7 @@ To download the Llama3.1-405B-Instruct model:
tune download meta-llama/Meta-Llama-3.1-405B-Instruct --ignore-patterns "original/consolidated*" --hf-token <HF_TOKEN>

To download the Llama3 weights of the above models, you can instead download from `Meta-Llama-3-8B-Instruct` and
`Meta-Llama-3-70B-Instruct`.
`Meta-Llama-3-70B-Instruct`, and remove the ignore patterns flag.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

unfortunately, 8b uses meta ckpt, and 70b uses hg ckpt. Ideally, we should update their configs to all use hf, so we can always ignore consolidated. To avoid adding too much info here, i suggested that the user can just remove the ignore patterns flag, downloading all models.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Remove the mention of this in kaggle_download as well.

@calvinpelletier
Copy link
Contributor

Not important but just for cleanliness, you could remove any --ignore-patterns None in the configs now that it is the default:

find /path/to/torchtune/recipes/configs -type f -name "*.yaml" -exec sed -i 's/ --ignore-patterns None//g' {} +

@felipemello1
Copy link
Contributor Author

Not important but just for cleanliness, you could remove any --ignore-patterns None in the configs now that it is the default:

find /path/to/torchtune/recipes/configs -type f -name "*.yaml" -exec sed -i 's/ --ignore-patterns None//g' {} +

yessir

@felipemello1 felipemello1 merged commit fff60d4 into pytorch:main Nov 21, 2024
17 checks passed
@felipemello1 felipemello1 deleted the remove_safetensor_default branch November 21, 2024 22:12
@ebsmothers ebsmothers mentioned this pull request Nov 26, 2024
44 tasks
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants