Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

i added hugging face read api key, but when i try to run a query i get a 503 #63

Closed
ralyodio opened this issue Apr 25, 2023 · 3 comments

Comments

@ralyodio
Copy link

i added hugging face read api key, but when i try to run a query i get a 503

@ralyodio
Copy link
Author

INFO:server.lib.inference:Requesting inference from databricks/dolly-v2-12b on huggingface
INFO:werkzeug:127.0.0.1 - - [24/Apr/2023 23:37:51] "POST /api/inference/text/stream HTTP/1.1" 200 -
ERROR:server.lib.inference:Error: Request failed: 503 Service Unavailable 

@ralyodio
Copy link
Author

how do i fix?

@zainhuda
Copy link
Collaborator

zainhuda commented May 1, 2023

seems like this is the huggingface API, you'll have to wait a bit for the model to "warm up" before you can call for inference. https://huggingface.co/docs/api-inference/faq#:~:text=Rate%20limits&text=We%20try%20to%20balance%20the,errors%20saying%20models%20are%20loading.

@zainhuda zainhuda closed this as completed May 1, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants