{ "Universal LLM Initialization": { "prefix": "ud-init", "body": [ "import { UniversalLLM } from 'universal-developer';", "", "const llm = new UniversalLLM({", " provider: '${1|anthropic,openai,qwen,gemini,ollama|}',", " apiKey: process.env.${2:${1/(anthropic|openai|qwen|gemini)/${1:/upcase}_API_KEY/}}", "});" ], "description": "Initialize a Universal Developer LLM instance" }, "Thinking Mode Generator": { "prefix": "ud-think", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/think ${3:What are the implications of ${4:technology} on ${5:domain}?}\"", "});" ], "description": "Generate response using thinking mode" }, "Fast Mode Generator": { "prefix": "ud-fast", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/fast ${3:${4:Summarize} ${5:this information}}\"", "});" ], "description": "Generate concise response using fast mode" }, "Loop Mode Generator": { "prefix": "ud-loop", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/loop --iterations=${3:3} ${4:Improve this ${5:text}: ${6:content}}\"", "});" ], "description": "Generate iteratively refined response using loop mode" }, "Reflection Mode Generator": { "prefix": "ud-reflect", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/reflect ${3:${4:Analyze} the ${5:implications} of ${6:topic}}\"", "});" ], "description": "Generate self-reflective response using reflection mode" }, "Fork Mode Generator": { "prefix": "ud-fork", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/fork --count=${3:2} ${4:Generate different ${5:approaches} to ${6:problem}}\"", "});" ], "description": "Generate multiple alternative responses using fork mode" }, "Chain Commands": { "prefix": "ud-chain", "body": [ "const response = await llm.generate({", " ${1:systemPrompt: `${2:You are a helpful assistant.}`,}", " prompt: \"/${3|think,loop,reflect,fork|} /${4|think,loop,reflect,fork|} ${5:Prompt text}\"", "});" ], "description": "Generate response using chained symbolic commands" }, "Custom Command Registration": { "prefix": "ud-custom", "body": [ "llm.registerCommand(\"${1:commandName}\", {", " description: \"${2:Command description}\",", " ${3:parameters: [", " {", " name: \"${4:paramName}\",", " description: \"${5:Parameter description}\",", " required: ${6:false},", " default: ${7:\"defaultValue\"}", " }", " ],}", " transform: async (prompt, options) => {", " ${8:// Custom implementation}", " const systemPrompt = `\\${options.systemPrompt || ''}", "${9:Custom system prompt instructions}`;", "", " return {", " systemPrompt,", " userPrompt: prompt,", " modelParameters: {", " ${10:temperature: 0.7}", " }", " };", " }", "});" ], "description": "Register a custom symbolic command" }, "Express API Integration": { "prefix": "ud-express", "body": [ "import express from 'express';", "import { UniversalLLM } from 'universal-developer';", "", "const app = express();", "app.use(express.json());", "", "const llm = new UniversalLLM({", " provider: '${1|anthropic,openai,qwen,gemini,ollama|}',", " apiKey: process.env.${2:${1/(anthropic|openai|qwen|gemini)/${1:/upcase}_API_KEY/}}", "});", "", "app.post('/api/generate', async (req, res) => {", " try {", " const { prompt, systemPrompt } = req.body;", " ", " // Get command from query param or default to /think", " const command = req.query.command || 'think';", " ", " const response = await llm.generate({", " systemPrompt,", " prompt: `/${command} ${prompt}`", " });", " ", " res.json({ response });", " } catch (error) {", " console.error('Error generating response:', error);", " res.status(500).json({ error: error.message });", " }", "});" ], "description": "Express API integration with Universal Developer" } }