-
Notifications
You must be signed in to change notification settings - Fork 195
worker: Add HF_TOKEN
to runner container on creation
#3693
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
Also provide clear hostname inside container when multiple containers are running on same host.
Codecov Report❌ Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## master #3693 +/- ##
===================================================
+ Coverage 31.92386% 31.92539% +0.00153%
===================================================
Files 156 156
Lines 47441 47445 +4
===================================================
+ Hits 15145 15147 +2
- Misses 31401 31402 +1
- Partials 895 896 +1
... and 4 files with indirect coverage changes Continue to review full report in Codecov by Sentry.
🚀 New features to boost your workflow:
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the PR @hjpotter92 Could you add in the description why we need to pass HF_TOKEN
into the runner? Why does runner need to connect to HuggingFace?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Added one suggestion
There's a possibility that this PR is not needed. @varshith15 mentioned getting rate limited. I'll add him as a reviewer as well and hold off on activity on this PR. |
so the main issue was that we are loading controlnets and unets in pytorch to detect the model type -- each model load takes about 3-4 sec and loading all controlnets adds upto the 15 sec josh noticed. this also adds to the vram as the pytorch models are not being unloaded after model type detection. ideally we shouldn't be loading any pytorch models and detect the model type from the model config instead (working on that). so we should pause the PR for now, can revisit if we still face the issue after i push the above fix @hjpotter92 |
Also provide clear hostname inside container when multiple containers are running on same host.