Skip to content

Support sharding in PyTorchHubMixin #2076

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
Wauplin opened this issue Mar 1, 2024 · 4 comments
Open

Support sharding in PyTorchHubMixin #2076

Wauplin opened this issue Mar 1, 2024 · 4 comments
Labels
mixins Everything related to ModelHuxMixin and cie

Comments

@Wauplin
Copy link
Contributor

Wauplin commented Mar 1, 2024

Depends on #2065.

@Wauplin Wauplin added the mixins Everything related to ModelHuxMixin and cie label Mar 1, 2024
@not-lain
Copy link
Contributor

not-lain commented Apr 7, 2024

@Wauplin
I think that #1995 by @NielsRogge has some interesting resources and maybe we can use that as a reference for this issue.

@Wauplin
Copy link
Contributor Author

Wauplin commented Apr 8, 2024

Actually #1995 was a start but I'd prefer to address #2065 first to make the logic reusable. I can take care of it.

Just to be sure in term of priorities, have we already encountered a library where we'd like to shard the weights or not yet?

@not-lain
Copy link
Contributor

not-lain commented Apr 8, 2024

@Wauplin not yet, all the AI models that i've worked on so far are small (<1GB) so all good so far

@Wauplin
Copy link
Contributor Author

Wauplin commented Apr 8, 2024

Thanks for confirming! Doesn't mean we shouldn't handle it but at least it's not critical :)

@Wauplin Wauplin added this to the in next release? milestone Apr 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
mixins Everything related to ModelHuxMixin and cie
Projects
None yet
Development

No branches or pull requests

2 participants