|
{ |
|
"Universal LLM Initialization": { |
|
"prefix": "ud-init", |
|
"body": [ |
|
"import { UniversalLLM } from 'universal-developer';", |
|
"", |
|
"const llm = new UniversalLLM({", |
|
" provider: '${1|anthropic,openai,qwen,gemini,ollama|}',", |
|
" apiKey: process.env.${2:${1/(anthropic|openai|qwen|gemini)/${1:/upcase}_API_KEY/}}", |
|
"});" |
|
], |
|
"description": "Initialize a Universal Developer LLM instance" |
|
}, |
|
"Thinking Mode Generator": { |
|
"prefix": "ud-think", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/think ${3:What are the implications of ${4:technology} on ${5:domain}?}\"", |
|
"});" |
|
], |
|
"description": "Generate response using thinking mode" |
|
}, |
|
"Fast Mode Generator": { |
|
"prefix": "ud-fast", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/fast ${3:${4:Summarize} ${5:this information}}\"", |
|
"});" |
|
], |
|
"description": "Generate concise response using fast mode" |
|
}, |
|
"Loop Mode Generator": { |
|
"prefix": "ud-loop", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/loop --iterations=${3:3} ${4:Improve this ${5:text}: ${6:content}}\"", |
|
"});" |
|
], |
|
"description": "Generate iteratively refined response using loop mode" |
|
}, |
|
"Reflection Mode Generator": { |
|
"prefix": "ud-reflect", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/reflect ${3:${4:Analyze} the ${5:implications} of ${6:topic}}\"", |
|
"});" |
|
], |
|
"description": "Generate self-reflective response using reflection mode" |
|
}, |
|
"Fork Mode Generator": { |
|
"prefix": "ud-fork", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/fork --count=${3:2} ${4:Generate different ${5:approaches} to ${6:problem}}\"", |
|
"});" |
|
], |
|
"description": "Generate multiple alternative responses using fork mode" |
|
}, |
|
"Chain Commands": { |
|
"prefix": "ud-chain", |
|
"body": [ |
|
"const response = await llm.generate({", |
|
" ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", |
|
" prompt: \"/${3|think,loop,reflect,fork|} /${4|think,loop,reflect,fork|} ${5:Prompt text}\"", |
|
"});" |
|
], |
|
"description": "Generate response using chained symbolic commands" |
|
}, |
|
"Custom Command Registration": { |
|
"prefix": "ud-custom", |
|
"body": [ |
|
"llm.registerCommand(\"${1:commandName}\", {", |
|
" description: \"${2:Command description}\",", |
|
" ${3:parameters: [", |
|
" {", |
|
" name: \"${4:paramName}\",", |
|
" description: \"${5:Parameter description}\",", |
|
" required: ${6:false},", |
|
" default: ${7:\"defaultValue\"}", |
|
" }", |
|
" ],}", |
|
" transform: async (prompt, options) => {", |
|
" ${8:// Custom implementation}", |
|
" const systemPrompt = `\\${options.systemPrompt || ''}", |
|
"${9:Custom system prompt instructions}`;", |
|
"", |
|
" return {", |
|
" systemPrompt,", |
|
" userPrompt: prompt,", |
|
" modelParameters: {", |
|
" ${10:temperature: 0.7}", |
|
" }", |
|
" };", |
|
" }", |
|
"});" |
|
], |
|
"description": "Register a custom symbolic command" |
|
}, |
|
"Express API Integration": { |
|
"prefix": "ud-express", |
|
"body": [ |
|
"import express from 'express';", |
|
"import { UniversalLLM } from 'universal-developer';", |
|
"", |
|
"const app = express();", |
|
"app.use(express.json());", |
|
"", |
|
"const llm = new UniversalLLM({", |
|
" provider: '${1|anthropic,openai,qwen,gemini,ollama|}',", |
|
" apiKey: process.env.${2:${1/(anthropic|openai|qwen|gemini)/${1:/upcase}_API_KEY/}}", |
|
"});", |
|
"", |
|
"app.post('/api/generate', async (req, res) => {", |
|
" try {", |
|
" const { prompt, systemPrompt } = req.body;", |
|
" ", |
|
" // Get command from query param or default to /think", |
|
" const command = req.query.command || 'think';", |
|
" ", |
|
" const response = await llm.generate({", |
|
" systemPrompt,", |
|
" prompt: `/${command} ${prompt}`", |
|
" });", |
|
" ", |
|
" res.json({ response });", |
|
" } catch (error) {", |
|
" console.error('Error generating response:', error);", |
|
" res.status(500).json({ error: error.message });", |
|
" }", |
|
"});" |
|
], |
|
"description": "Express API integration with Universal Developer" |
|
} |
|
} |
|
|