[FEEDBACK] Inference Providers
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
Love that I can call DeepSeek R1 directly from the Hub ๐ฅ
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="together",
api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)
messages = [
{
"role": "user",
"content": "What is the capital of France?"
}
]
completion = client.chat.completions.create(
model="deepseek-ai/DeepSeek-R1",
messages=messages,
max_tokens=500
)
print(completion.choices[0].message)
Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
Thanks for your quick reply, good to know!
Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...
Could be good to add featherless.ai
TitanML !!
let me run the numbers, but for HF-inference the change should not be that drastic (most HF-inference requests are priced very very cheaply especially of course CPU based models)
The 20,000 daily figure was a bit unrealistic given it was "best-effort" meaning the rate of failures was quite high. Note that from now on, we exclude failing requests from those counts (we didn't up till now)
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
I tested it and I love it, super easy!
The first question from my company was:
With inference providers can we setup something like private endpoints?
@Moibe do you use HF-inference, or external providers?
@julien-c yes I was using the hf-inference and usage was deducted from the 20,000. And about other providers, it allowed me to use them totally free (without substracting from the 20,000) and with a limited quota, meaning that I was able to make some usage of fal-ai until it messages that quota was excedded, then I used together, sambanova, etc. It was great beause it allowed me to test all provider even without having an account. I understand the change, and I think prices are still reasonable. I was just a shocking change, but I'm fine. I even got an account on fal.ai after the free test and I'm satisfied with the service amd the ease of use all from hf interface.
Let's add nineteen.ai PLEAAASE
With inference providers can we setup something like private endpoints?
@levalencia i don't think so but have you looked into Inference Endpoints? (it's dedicated instances)
THE HF INT API IS DENYING MY SERVICE BARKIN' "FAILED TO FETCH" (ILLEGAL EXCUSE!!!) (USED 0.20 OUT 2$ AS PRO)!!! SUPPORT, END THIS MADNESS NOW!!! TOMORROW, COURT AND PRESS, IF DON'T!!!
Subject: Add new provider to the Inference Endpoint.
I work at https://docs.predictionguard.com/home/getting-started/welcome and would like to explore if it is possible to add Prediction Guard to the list of inference providers on Hugging Face. Let me know how to go about this.