Spaces:
Running
Running
docs(README): recent section links (#1616)
Browse filesCo-authored-by: Nathan Sarrazin <[email protected]>
README.md
CHANGED
@@ -107,7 +107,7 @@ If you don't want to configure, setup, and launch your own Chat UI yourself, you
|
|
107 |
|
108 |
You can deploy your own customized Chat UI instance with any supported [LLM](https://huggingface.co/models?pipeline_tag=text-generation&sort=trending) of your choice on [Hugging Face Spaces](https://huggingface.co/spaces). To do so, use the chat-ui template [available here](https://huggingface.co/new-space?template=huggingchat/chat-ui-template).
|
109 |
|
110 |
-
Set `HF_TOKEN` in [Space secrets](https://huggingface.co/docs/hub/spaces-overview#managing-secrets
|
111 |
|
112 |
Read the full tutorial [here](https://huggingface.co/docs/hub/spaces-sdks-docker-chatui#chatui-on-spaces).
|
113 |
|
@@ -316,7 +316,7 @@ The following is the default `chatPromptTemplate`, although newlines and indenti
|
|
316 |
|
317 |
#### Multi modal model
|
318 |
|
319 |
-
We currently support [IDEFICS](https://huggingface.co/blog/idefics) (hosted on TGI), OpenAI and Claude 3 as multimodal models. You can enable it by setting `multimodal: true` in your `MODELS` configuration. For IDEFICS, you must have a [PRO HF Api token](https://huggingface.co/settings/tokens). For OpenAI, see the [OpenAI section](#
|
320 |
|
321 |
```env
|
322 |
{
|
@@ -695,7 +695,7 @@ You can get the `accessKey` and `secretKey` from your AWS user, under programmat
|
|
695 |
|
696 |
You can also use Cloudflare Workers AI to run your own models with serverless inference.
|
697 |
|
698 |
-
You will need to have a Cloudflare account, then get your [account ID](https://developers.cloudflare.com/fundamentals/setup/find-account-and-zone-ids/) as well as your [API token](https://developers.cloudflare.com/workers-ai/get-started/rest-api/#1-get-
|
699 |
|
700 |
You can either specify them directly in your `.env.local` using the `CLOUDFLARE_ACCOUNT_ID` and `CLOUDFLARE_API_TOKEN` variables, or you can set them directly in the endpoint config.
|
701 |
|
|
|
107 |
|
108 |
You can deploy your own customized Chat UI instance with any supported [LLM](https://huggingface.co/models?pipeline_tag=text-generation&sort=trending) of your choice on [Hugging Face Spaces](https://huggingface.co/spaces). To do so, use the chat-ui template [available here](https://huggingface.co/new-space?template=huggingchat/chat-ui-template).
|
109 |
|
110 |
+
Set `HF_TOKEN` in [Space secrets](https://huggingface.co/docs/hub/spaces-overview#managing-secrets) to deploy a model with gated access or a model in a private repository. It's also compatible with [Inference for PROs](https://huggingface.co/blog/inference-pro) curated list of powerful models with higher rate limits. Make sure to create your personal token first in your [User Access Tokens settings](https://huggingface.co/settings/tokens).
|
111 |
|
112 |
Read the full tutorial [here](https://huggingface.co/docs/hub/spaces-sdks-docker-chatui#chatui-on-spaces).
|
113 |
|
|
|
316 |
|
317 |
#### Multi modal model
|
318 |
|
319 |
+
We currently support [IDEFICS](https://huggingface.co/blog/idefics) (hosted on TGI), OpenAI and Claude 3 as multimodal models. You can enable it by setting `multimodal: true` in your `MODELS` configuration. For IDEFICS, you must have a [PRO HF Api token](https://huggingface.co/settings/tokens). For OpenAI, see the [OpenAI section](#openai-api-compatible-models). For Anthropic, see the [Anthropic section](#anthropic).
|
320 |
|
321 |
```env
|
322 |
{
|
|
|
695 |
|
696 |
You can also use Cloudflare Workers AI to run your own models with serverless inference.
|
697 |
|
698 |
+
You will need to have a Cloudflare account, then get your [account ID](https://developers.cloudflare.com/fundamentals/setup/find-account-and-zone-ids/) as well as your [API token](https://developers.cloudflare.com/workers-ai/get-started/rest-api/#1-get-api-token-and-account-id) for Workers AI.
|
699 |
|
700 |
You can either specify them directly in your `.env.local` using the `CLOUDFLARE_ACCOUNT_ID` and `CLOUDFLARE_API_TOKEN` variables, or you can set them directly in the endpoint config.
|
701 |
|