[FEEDBACK] Inference Providers
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
Love that I can call DeepSeek R1 directly from the Hub 🔥
from huggingface_hub import InferenceClient client = InferenceClient( provider="together", api_key="xxxxxxxxxxxxxxxxxxxxxxxx" ) messages = [ { "role": "user", "content": "What is the capital of France?" } ] completion = client.chat.completions.create( model="deepseek-ai/DeepSeek-R1", messages=messages, max_tokens=500 ) print(completion.choices[0].message)
Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
Thanks for your quick reply, good to know!
Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...
Could be good to add featherless.ai
TitanML !!
Dear Hugging Face community,
I'm reaching out from Runware. We provide a high-quality media through a fast, affordable API. We are excited in becoming an official inference provider on Hugging Face Hub.
We are already on the team plan. But still unable to post model mapping as we keep on getting Cannot POST /api/partners/runware/models
error.
Kindly let us know the next step to move forward.
Best,
Sirshendu
email: sirshendu.ganguly@runware.ai
Hey there all,
We're Bytez, and we're the largest inference provider on the internet! (We offer inference for 170k+ models.)
We stepped through the integration guide and upgraded to a team account, but are encountering the same issue @sirshendu1997 is encountering.
We also reached out to enterprise-customer@huggingface.co but have not received word back after a few days, so we're trying this avenue.
Please let us know what we need to do to get our integration working. We'd love to be a provider for HF!
Hi,
I'm Charles from Boost Run. We own our compute and operate our data centers, running large clusters equipped with Blackwell, Hopper, and other model GPUs.
Back in 2022-24, I was one of the early engineers on Amazon Bedrock. At the time, the push for dedicated GenAI inference providers was still fairly new. I had a front-row seat to what worked, and what became fragile as things scaled. Our goal is to build a lean and efficient machine, processing the most important workloads in a secure and highly available manner.
I'll be going through the onboarding docs over the coming weeks and am excited to get more familiar with everything. Hopefully this can be the beginning of a strong partnership.
PS. AI is a better writer than myself 😀
Also feel free to reach out to me cr@boostrun.com or can invite to our slack org
Hi Hugging Face team, 👋
I’m with GPT Porto (https://www.gptproto.com/), an AI API platform focused on providing safe, stable, fast, and affordable inference for developers and enterprises. With just one API key, our users can access most mainstream models, including Hugging Face models, while enjoying reliable infrastructure and cost efficiency.
We’ve seen strong adoption from teams who value predictable performance, security, and competitive pricing for both experimentation and production workloads. Many of them already integrate Hugging Face models through GPT Proto to streamline deployment and reduce costs.
We’d love to explore becoming an official Inference Provider on Hugging Face, so that more builders in your ecosystem can benefit from a secure, high-performance, and budget-friendly option for model inference.
Looking forward to collaborating!
Contact us: gptproto@venuslondontechnology.co.uk
Best regards,
Team GPT Proto
Hi Julien would I be able to reach out to you or someone from HF about the API Key management and Billing?
https://huggingface.co/docs/inference-providers/en/register-as-a-provider#4-billing
Thanks,
Boost Run
Prev: https://huggingface.co/spaces/huggingface/HuggingDiscussions/discussions/49#68d564aa7a7e392a927c11f6
Would be great to have Simplismart to the list!