Cole Medin commited on
Commit
8c4933d
Β·
1 Parent(s): d7bda6f

Updated README with new providers and a running list of features to add to the fork

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -2,7 +2,23 @@
2
 
3
  # Bolt.new Fork by Cole Medin
4
 
5
- This fork of bolt.new allows you to choose the LLM that you use for each prompt! Currently you can use OpenAI, Anthropic, Ollama, or Groq models - and it is easily extended to use any other model supported by the Vercel AI SDK! See instructions below for running this locally and extending to include more models.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
  # Bolt.new: AI-Powered Full-Stack Web Development in the Browser
8
 
 
2
 
3
  # Bolt.new Fork by Cole Medin
4
 
5
+ This fork of bolt.new allows you to choose the LLM that you use for each prompt! Currently you can use OpenAI, Anthropic, Ollama, OpenRouter, Gemini, or Groq models - and it is easily extended to use any other model supported by the Vercel AI SDK! See instructions below for running this locally and extending to include more models.
6
+
7
+ # Requested Additions to this Fork - Feel Free to Contribute!!
8
+
9
+ - βœ… OpenRouter Integration (@coleam00)
10
+ - βœ… Gemini Integration (@jonathands)
11
+ - βœ… Autogenerate Ollama models from what is downloaded (@mosquet)
12
+ - βœ… Filter models by provider (@jasonm23)
13
+ - βœ… Download project as ZIP (@fabwaseem)
14
+ - ⬜ LM Studio Integration
15
+ - ⬜ DeepSeek API Integration
16
+ - ⬜ Better prompting for smaller LLMs (code window sometimes doesn't start)
17
+ - ⬜ Attach images to prompts
18
+ - ⬜ Run agents in the backend instead of a single model call
19
+ - ⬜ Publish projects directly to GitHub
20
+ - ⬜ Load local projects into the app
21
+ - ⬜ Improvements to the main Bolt.new prompt in `app\lib\.server\llm\prompts.ts` (there is definitely opportunity there)
22
 
23
  # Bolt.new: AI-Powered Full-Stack Web Development in the Browser
24