[FEEDBACK] Inference Providers
Any inference provider you love, and that you'd like to be able to access directly from the Hub?
Love that I can call DeepSeek R1 directly from the Hub π₯
from huggingface_hub import InferenceClient
client = InferenceClient(
provider="together",
api_key="xxxxxxxxxxxxxxxxxxxxxxxx"
)
messages = [
{
"role": "user",
"content": "What is the capital of France?"
}
]
completion = client.chat.completions.create(
model="deepseek-ai/DeepSeek-R1",
messages=messages,
max_tokens=500
)
print(completion.choices[0].message)
Is it possible to set a monthly payment budget or rate limits for all the external providers? I don't see such options in billings tab. In case a key is or session token is stolen, it can be quite dangerous to my thin wallet:(
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
@benhaotang you already get spending notifications when crossing important thresholds ($10, $100, $1,000) but we'll add spending limits in the future
Thanks for your quick reply, good to know!
Would be great if you could add Nebius AI Studio to the list :) New inference provider on the market, with the absolute cheapest prices and the highest rate limits...
Could be good to add featherless.ai
TitanML !!
The message for any HFInference is: "Our latest automated health check on this model for this provider did not complete successfully." Is this temporary or HFInference won't process certain models anymore, or maybe it is a bug?
Dear Hugging Face Team,
Greetings from Cyfuture AI!
We are reaching out to explore a potential collaboration with Hugging Face. As a rapidly growing enterprise-grade AI solutions provider, Cyfuture AI offers robust and scalable inference capabilities powered by high-performance GPU infrastructure.
We would be keen to join Hugging Face as an official inference provider to support model deployment and inference workloads for your global community. We believe this integration would bring mutual valueβenhancing access to affordable, high-speed inference while expanding our reach within the AI ecosystem.
Please let us know the next steps or any prerequisites required to move forward with this partnership.
Looking forward to your response.
Regards,
Cyfuture.ai
Email us at - [email protected]
Hi Hugging Face Team,
We are from GmiCloud (https://inference-engine.gmicloud.ai). We want to be an inference provider on Hugging Face. At gmicloud, we focus on LLM inference optimizations. We started to follow the instruction at https://huggingface.co/docs/inference-providers/register-as-a-provider#register-the-provider. While it needs to reach out first. Weβd greatly appreciate any guidance or support from the community on how to move forward with becoming an official inference provider on the platform.
Thanks in advance!
Gmi Cloud AI
Email us at - [email protected]
Dear Hugging Face Team,
We're reaching out from Swarmind.ai, a high-performance AI infrastructure company, to express interest in becoming an official inference provider on your platform.
We offer scalable, GPU-powered inference optimized for production workloads, and believe this integration would benefit both communities.
Let us know the next steps to move forward.
Best,
Swarmind Team
[email protected]
Dear Hugging Face Team,
We're reaching out from metay.ai , a high performance AI infra company. We are super excited in becoming an official inference provider.
Pls let us know the next step to move forward.
Best,
Sam
my company email: [email protected]
CometAPI - Hugging Face Inference Provider Application
Dear Hugging Face Team,
We're reaching out from CometAPI to express our strong interest in becoming an official inference provider on the Hugging Face platform.
About CometAPI
CometAPI is a comprehensive AI platform that provides unified access to over 500 cutting-edge AI models through a single, powerful API. We serve as a bridge between developers and the latest AI capabilities, offering seamless integration with minimal complexity.
Key Highlights:
- π 500+ AI Models: Extensive collection including GPT-5, Claude, GLM-4.5, Qwen3-Coder, Kimi K2, Grok 4, and many more
- π¨ Multi-Modal Support: Text generation, image creation, video production, music composition, and audio processing
- β‘ Unified API: Single endpoint for all AI services with OpenAI-compatible API format
- π Real-time Updates: Instant access to newly released AI models
- π° Competitive Pricing: Cost-effective solutions with flexible usage plans
- π‘οΈ Enterprise Ready: Robust infrastructure with enterprise-grade security and reliability
Why Partner with CometAPI?
Comprehensive Model Coverage: We provide access to the industry's most extensive collection of AI models, including many that are not available through other providers.
OpenAI API Compatibility: Our API strictly follows OpenAI specifications for LLMs and VLMs, making integration seamless and straightforward.
High-Performance Infrastructure: Built on scalable, enterprise-grade infrastructure designed to handle production workloads with low latency.
Global Reach: Serving developers worldwide with reliable service and comprehensive documentation.
Active Community: Growing ecosystem with active Discord community and comprehensive developer support.
Technical Capabilities
- API Compatibility: Full OpenAI API compliance for chat completions and text generation
- Supported Tasks:
- Conversational AI (text-generation, image-text-to-text)
- Text generation
- Multi-modal capabilities
- Authentication: Bearer token authentication system
- Billing Integration: Ready to implement request ID tracking and cost reporting API
- Rate Limiting: Robust rate limiting and quota management systems
Business Information
- Company: CometAPI
- Website: www.cometapi.com
- API Documentation: api.cometapi.com/doc
- Console: api.cometapi.com/console
Contact Information
Primary Contact: [email protected]
Additional Resources:
- Discord Community: Join our Discord
- Twitter: @cometapi2025
- GitHub: github.com/CometAPI
- Support Email: [email protected]
Next Steps
We are fully committed to following the complete integration process outlined in your documentation:
- β Prerequisites: We have OpenAI-compatible API and understand the task requirements
- π JS Client Integration: Ready to submit PR to huggingface.js repository
- π Model Mapping API: Ready to register our model mappings once enabled
- π Billing Implementation: Can implement the required billing endpoint API
- π Python Client Integration: Ready to submit PR to huggingface_hub
- π Documentation: Ready to create comprehensive provider documentation
Value Proposition for Hugging Face Users
By integrating CometAPI as an inference provider, Hugging Face users will gain:
- Access to 500+ AI models through a single provider
- Competitive pricing with transparent cost structure
- High reliability and low-latency inference
- Regular updates with the latest model releases
- Comprehensive multi-modal capabilities
- Enterprise-grade support and SLA commitments
We believe this partnership will bring significant value to the Hugging Face community by providing broader model access and enhanced AI capabilities.
We would greatly appreciate the opportunity to discuss this partnership further and are ready to begin the technical integration process immediately upon approval.
Thank you for considering our application. We look forward to your response and the possibility of collaborating with the Hugging Face team.
Best regards,
CometAPI Development Team
[email protected]
www.cometapi.com
Ready to revolutionize AI accessibility together? Let's make it happen! π