Update tabbed.py
Browse files
tabbed.py
CHANGED
@@ -90,14 +90,8 @@ with gr.Blocks() as demo:
|
|
90 |
with gr.Row():
|
91 |
with gr.Column():
|
92 |
gr.Markdown(f"""
|
93 |
-
###
|
94 |
- This is the [{config["hub"]["repo_id"]}](https://huggingface.co/{config["hub"]["repo_id"]}) model file [{config["hub"]["filename"]}](https://huggingface.co/{config["hub"]["repo_id"]}/blob/main/{config["hub"]["filename"]})
|
95 |
-
- This Space uses GGML with GPU support, so it can quickly run larger models on smaller GPUs & VRAM.
|
96 |
-
- This is running on a smaller, shared GPU, so it may take a few seconds to respond.
|
97 |
-
- [Duplicate the Space](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui?duplicate=true) to skip the queue and run in a private space or to use your own GGML models.
|
98 |
-
- When using your own models, simply update the [config.yml](https://huggingface.co/spaces/openaccess-ai-collective/ggml-ui/blob/main/config.yml)
|
99 |
-
- Contribute at [https://github.com/OpenAccess-AI-Collective/ggml-webui](https://github.com/OpenAccess-AI-Collective/ggml-webui)
|
100 |
-
- Many thanks to [TheBloke](https://huggingface.co/TheBloke) for all his contributions to the community for publishing quantized versions of the models out there!
|
101 |
""")
|
102 |
with gr.Tab("Instruct"):
|
103 |
gr.Markdown("# GGML Spaces Instruct Demo")
|
|
|
90 |
with gr.Row():
|
91 |
with gr.Column():
|
92 |
gr.Markdown(f"""
|
93 |
+
### One
|
94 |
- This is the [{config["hub"]["repo_id"]}](https://huggingface.co/{config["hub"]["repo_id"]}) model file [{config["hub"]["filename"]}](https://huggingface.co/{config["hub"]["repo_id"]}/blob/main/{config["hub"]["filename"]})
|
|
|
|
|
|
|
|
|
|
|
|
|
95 |
""")
|
96 |
with gr.Tab("Instruct"):
|
97 |
gr.Markdown("# GGML Spaces Instruct Demo")
|