Spaces:
Running
Running
File size: 39,032 Bytes
ed8e0ee 9df5dcc ed8e0ee 713f2f6 9df5dcc 2cae477 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc ed8e0ee 2cae477 9df5dcc 2cae477 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 e6d9a41 713f2f6 e6d9a41 713f2f6 e6d9a41 713f2f6 e6d9a41 713f2f6 e6d9a41 9df5dcc 713f2f6 9df5dcc e6d9a41 9df5dcc e6d9a41 713f2f6 e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc e6d9a41 9df5dcc 713f2f6 9df5dcc e6d9a41 713f2f6 e6d9a41 9df5dcc 713f2f6 e6d9a41 713f2f6 9df5dcc e6d9a41 713f2f6 ed8e0ee 9df5dcc ed8e0ee 9df5dcc 2cae477 713f2f6 2cae477 713f2f6 9df5dcc 2cae477 9df5dcc 2cae477 713f2f6 2cae477 9df5dcc 2cae477 713f2f6 9df5dcc 2cae477 9df5dcc 2cae477 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc 713f2f6 9df5dcc e6d9a41 9df5dcc e6d9a41 713f2f6 e6d9a41 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 e6d9a41 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc e6d9a41 9df5dcc e6d9a41 9df5dcc e6d9a41 9df5dcc e6d9a41 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 e6d9a41 713f2f6 e6d9a41 713f2f6 e6d9a41 713f2f6 9df5dcc e6d9a41 713f2f6 e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc ed8e0ee 9df5dcc ed8e0ee 713f2f6 ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 713f2f6 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc e6d9a41 713f2f6 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc e6d9a41 9df5dcc 713f2f6 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc ed8e0ee 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc ed8e0ee 9df5dcc e6d9a41 9df5dcc e6d9a41 9df5dcc e6d9a41 9df5dcc e6d9a41 ed8e0ee 9df5dcc 713f2f6 9df5dcc 713f2f6 e6d9a41 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc ed8e0ee 9df5dcc ed8e0ee 713f2f6 ed8e0ee 713f2f6 ed8e0ee 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc e6d9a41 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 ed8e0ee 9df5dcc ed8e0ee 9df5dcc 713f2f6 9df5dcc 713f2f6 ed8e0ee 9df5dcc 2cae477 713f2f6 9df5dcc 713f2f6 9df5dcc 713f2f6 e6d9a41 713f2f6 9df5dcc 713f2f6 ed8e0ee 9df5dcc ed8e0ee 9df5dcc 2cae477 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 |
import express from 'express';
// Import the 'fal' object directly for configuration within the retry loop
import { fal } from '@fal-ai/client';
// --- Key Management Setup ---
// Read comma-separated keys from the SINGLE environment variable FAL_KEY
const FAL_KEY_STRING = process.env.FAL_KEY;
// Read the custom API Key for proxy authentication
const API_KEY = process.env.API_KEY;
if (!FAL_KEY_STRING) {
console.error("ERROR: FAL_KEY environment variable is not set.");
console.error("Ensure FAL_KEY contains a comma-separated list of your Fal AI keys.");
process.exit(1); // Exit if no Fal keys are provided
}
// Parse the comma-separated keys from FAL_KEY_STRING
const falKeys = FAL_KEY_STRING.split(',')
.map(key => key.trim()) // Remove leading/trailing whitespace
.filter(key => key.length > 0); // Remove any empty strings resulting from extra commas
if (falKeys.length === 0) {
console.error("ERROR: No valid Fal keys found in the FAL_KEY environment variable after parsing.");
console.error("Ensure FAL_KEY is a comma-separated list, e.g., 'key1,key2,key3'.");
process.exit(1); // Exit if parsing results in zero valid keys
}
if (!API_KEY) {
console.error("ERROR: API_KEY environment variable is not set.");
process.exit(1); // Exit if the proxy auth key is missing
}
let currentKeyIndex = 0;
// Keep track of keys that failed persistently during runtime
const invalidKeys = new Set();
console.log(`Loaded ${falKeys.length} Fal AI Key(s) from the FAL_KEY environment variable.`);
/**
* Gets the next available valid Fal AI key in a round-robin fashion.
* Skips keys that have been marked as invalid.
* @returns {object|null} An object containing the key and its original index { key, index }, or null if no valid keys remain.
*/
function getNextValidKey() {
// Check if all keys have been marked as invalid
if (invalidKeys.size >= falKeys.length) {
console.error("All Fal AI keys are marked as invalid.");
return null; // No valid keys left
}
const initialIndex = currentKeyIndex;
let attempts = 0; // Prevent infinite loops in edge cases
while (attempts < falKeys.length) {
const keyIndex = currentKeyIndex % falKeys.length;
const key = falKeys[keyIndex];
// Move to the next index for the *next* call, regardless of validity
currentKeyIndex = (keyIndex + 1) % falKeys.length;
// Check if the current key is NOT in the invalid set
if (!invalidKeys.has(key)) {
// Found a valid key
console.log(`Using Fal Key index: ${keyIndex} (from FAL_KEY list)`);
return { key, index: keyIndex }; // Return the key and its original index
} else {
console.log(`Skipping invalid Fal Key index: ${keyIndex}`);
}
attempts++;
// Safety check: If we've looped back to the start after trying, break.
// This is mostly redundant due to the invalidKeys.size check, but acts as a safeguard.
if (currentKeyIndex === initialIndex && attempts > 0) {
console.warn("Looped through all keys, potentially all are invalid.");
break;
}
}
// If we exit the loop, it means no valid key was found
console.error("Could not find a valid Fal AI key after checking all potentially available keys.");
return null;
}
/**
* Checks if an error object likely indicates an issue with the Fal AI API key.
* This is heuristic-based and may need refinement based on observed Fal errors.
* @param {Error|object} error - The error object caught.
* @returns {boolean} True if the error seems key-related, false otherwise.
*/
function isKeyRelatedError(error) {
if (!error) return false; // Handle null/undefined errors
const message = error.message?.toLowerCase() || '';
// Check common HTTP status properties
const status = error.status || error.statusCode;
// Check for specific HTTP status codes (401: Unauthorized, 403: Forbidden, 429: Too Many Requests/Quota)
if (status === 401 || status === 403 || status === 429) {
console.warn(`Detected potential key-related error (HTTP Status: ${status}).`);
return true;
}
// Check for common error message patterns (case-insensitive)
const keyErrorPatterns = [
'invalid api key', 'authentication failed', 'permission denied',
'quota exceeded', 'forbidden', 'unauthorized', 'rate limit',
'credentials', 'api key missing', 'invalid credential'
];
if (keyErrorPatterns.some(pattern => message.includes(pattern))) {
console.warn(`Detected potential key-related error (message contains relevant pattern: "${message}")`);
return true;
}
// Add more specific checks based on observed Fal AI errors if needed
// e.g., if (error.code === 'FAL_AUTH_FAILURE') return true;
return false;
}
// --- End Key Management Setup ---
const app = express();
// Increase payload size limits if needed
app.use(express.json({ limit: '50mb' }));
app.use(express.urlencoded({ extended: true, limit: '50mb' }));
const PORT = process.env.PORT || 3000;
// API Key Authentication Middleware
const apiKeyAuth = (req, res, next) => {
const authHeader = req.headers['authorization'];
if (!authHeader) {
console.warn('Unauthorized: No Authorization header provided');
return res.status(401).json({ error: 'Unauthorized: No API Key provided' });
}
// Expecting "Bearer YOUR_API_KEY"
const authParts = authHeader.split(' ');
if (authParts.length !== 2 || authParts[0].toLowerCase() !== 'bearer') {
console.warn('Unauthorized: Invalid Authorization header format. Expected "Bearer <key>".');
return res.status(401).json({ error: 'Unauthorized: Invalid Authorization header format' });
}
const providedKey = authParts[1];
if (providedKey !== API_KEY) {
console.warn('Unauthorized: Invalid API Key provided.');
return res.status(401).json({ error: 'Unauthorized: Invalid API Key' });
}
// Key is valid, proceed to the next middleware or route handler
next();
};
// Apply API Key Authentication to relevant endpoints
app.use(['/v1/models', '/v1/chat/completions'], apiKeyAuth);
// === Global Limits Definition ===
const PROMPT_LIMIT = 4800; // Max length for the main 'prompt' field
const SYSTEM_PROMPT_LIMIT = 4800; // Max length for the 'system_prompt' field
// === End Limits Definition ===
// Define the list of models supported by fal-ai/any-llm (Update as needed)
const FAL_SUPPORTED_MODELS = [
"anthropic/claude-3.7-sonnet",
"anthropic/claude-3.5-sonnet",
"anthropic/claude-3-5-haiku",
"anthropic/claude-3-haiku",
"google/gemini-pro-1.5",
"google/gemini-flash-1.5",
"google/gemini-flash-1.5-8b",
"google/gemini-2.0-flash-001",
"meta-llama/llama-3.2-1b-instruct",
"meta-llama/llama-3.2-3b-instruct",
"meta-llama/llama-3.1-8b-instruct",
"meta-llama/llama-3.1-70b-instruct",
"openai/gpt-4o-mini",
"openai/gpt-4o",
"deepseek/deepseek-r1",
"meta-llama/llama-4-maverick",
"meta-llama/llama-4-scout"
// Add or remove models here
];
// Helper function to extract the owner/organization from a model ID string
const getOwner = (modelId) => {
if (modelId && typeof modelId === 'string' && modelId.includes('/')) {
return modelId.split('/')[0];
}
// Default owner if format is unexpected or missing
return 'fal-ai';
}
// GET /v1/models endpoint - Returns the list of supported models
app.get('/v1/models', (req, res) => {
console.log("Received request for GET /v1/models");
try {
const modelsData = FAL_SUPPORTED_MODELS.map(modelId => ({
id: modelId,
object: "model",
created: Math.floor(Date.now() / 1000), // Use current timestamp
owned_by: getOwner(modelId)
}));
res.json({ object: "list", data: modelsData });
console.log("Successfully returned model list.");
} catch (error) {
console.error("Error processing GET /v1/models:", error);
res.status(500).json({ error: "Failed to retrieve model list." });
}
});
/**
* Converts OpenAI-style messages array to Fal AI's prompt and system_prompt format.
* Implements System prompt top-priority, separator, and recency-based history filling.
* Includes robustness checks for input validation and internal errors.
* @param {Array<object>} messages - Array of message objects ({ role: string, content: string }).
* @returns {object} An object containing { system_prompt: string, prompt: string }.
* @throws {Error} If input is invalid or an internal processing error occurs.
*/
function convertMessagesToFalPrompt(messages) {
// console.log("Entering convertMessagesToFalPrompt with messages:", JSON.stringify(messages, null, 2)); // Debug log
// --- Input Validation ---
if (!Array.isArray(messages)) {
console.error("Error in convertMessagesToFalPrompt: Input 'messages' is not an array.");
throw new Error("Invalid input: 'messages' must be an array.");
}
if (messages.length === 0) {
console.warn("Warning in convertMessagesToFalPrompt: Input 'messages' array is empty.");
return { system_prompt: "", prompt: "" }; // Return empty if no messages
}
// --- End Input Validation ---
try { // Wrap main logic in try...catch for internal errors
let fixed_system_prompt_content = "";
const conversation_message_blocks = [];
// console.log(`Original messages count: ${messages.length}`);
// 1. Separate System messages, format User/Assistant messages
for (const message of messages) {
// Validate individual message structure
if (!message || typeof message !== 'object' || typeof message.role !== 'string') {
console.warn(`Skipping invalid message object in convertMessagesToFalPrompt: ${JSON.stringify(message)}`);
continue; // Skip malformed message
}
// Safely handle content (null/undefined become empty string)
let content = (message.content === null || message.content === undefined) ? "" : String(message.content);
switch (message.role) {
case 'system':
// Append all system messages together
fixed_system_prompt_content += `System: ${content}\n\n`;
break;
case 'user':
conversation_message_blocks.push(`Human: ${content}\n\n`);
break;
case 'assistant':
conversation_message_blocks.push(`Assistant: ${content}\n\n`);
break;
default:
// Log unsupported roles but continue processing
console.warn(`Unsupported role encountered in convertMessagesToFalPrompt: ${message.role}. Skipping message.`);
continue;
}
}
// 2. Truncate combined system messages if they exceed the limit
if (fixed_system_prompt_content.length > SYSTEM_PROMPT_LIMIT) {
const originalLength = fixed_system_prompt_content.length;
fixed_system_prompt_content = fixed_system_prompt_content.substring(0, SYSTEM_PROMPT_LIMIT);
console.warn(`Combined system messages truncated from ${originalLength} to ${SYSTEM_PROMPT_LIMIT} characters.`);
}
// Trim trailing whitespace from the fixed system content
fixed_system_prompt_content = fixed_system_prompt_content.trim();
// 3. Calculate remaining space in system_prompt for history
// Consider potential separator length later if needed
let space_occupied_by_fixed_system = 0;
if (fixed_system_prompt_content.length > 0) {
// Approximate space: content + potential separator overhead (\n\n...\n\n)
space_occupied_by_fixed_system = fixed_system_prompt_content.length + 4; // Heuristic for spacing
}
const remaining_system_limit = Math.max(0, SYSTEM_PROMPT_LIMIT - space_occupied_by_fixed_system);
// console.log(`Trimmed fixed system prompt length: ${fixed_system_prompt_content.length}. Approx remaining system history limit: ${remaining_system_limit}`);
// 4. Fill history backwards (recency): Prioritize 'prompt', then 'system_prompt' overflow
const prompt_history_blocks = []; // For the main 'prompt' field
const system_prompt_history_blocks = []; // For history overflowing into 'system_prompt'
let current_prompt_length = 0;
let current_system_history_length = 0;
let promptFull = (PROMPT_LIMIT <= 0); // Handle zero limit case
let systemHistoryFull = (remaining_system_limit <= 0);
// console.log(`Processing ${conversation_message_blocks.length} user/assistant messages for recency filling.`);
for (let i = conversation_message_blocks.length - 1; i >= 0; i--) {
const message_block = conversation_message_blocks[i];
// Ensure message_block is a string before getting length
const block_length = (typeof message_block === 'string') ? message_block.length : 0;
if (block_length === 0) continue; // Skip empty blocks
// If both slots are full, stop processing older messages
if (promptFull && systemHistoryFull) {
// console.log(`Both prompt and system history slots full. Omitting older messages from index ${i}.`);
break;
}
// Try fitting into the main 'prompt' first
if (!promptFull) {
if (current_prompt_length + block_length <= PROMPT_LIMIT) {
prompt_history_blocks.unshift(message_block); // Add to beginning
current_prompt_length += block_length;
continue; // Message placed, move to next older message
} else {
promptFull = true; // Main prompt is now full
// console.log(`Prompt limit (${PROMPT_LIMIT}) reached. Trying system history slot.`);
}
}
// If prompt is full, try fitting into the 'system_prompt' remaining space
if (!systemHistoryFull) {
if (current_system_history_length + block_length <= remaining_system_limit) {
system_prompt_history_blocks.unshift(message_block); // Add to beginning
current_system_history_length += block_length;
continue; // Message placed, move to next older message
} else {
systemHistoryFull = true; // System history slot is now full
// console.log(`System history limit (${remaining_system_limit}) reached.`);
}
}
}
// 5. Combine the final prompt and system_prompt parts
const system_prompt_history_content = system_prompt_history_blocks.join('').trim();
const final_prompt = prompt_history_blocks.join('').trim(); // Main prompt content
// Separator to distinguish fixed system prompt from overflow history
const SEPARATOR = "\n\n------- Earlier Conversation History -------\n\n";
let final_system_prompt = "";
// Check if we have content for each part
const hasFixedSystem = fixed_system_prompt_content.length > 0;
const hasSystemHistory = system_prompt_history_content.length > 0;
if (hasFixedSystem && hasSystemHistory) {
// Both parts exist: Combine with separator
final_system_prompt = fixed_system_prompt_content + SEPARATOR + system_prompt_history_content;
// console.log("Combining fixed system prompt and history with separator.");
} else if (hasFixedSystem) {
// Only fixed system prompt exists
final_system_prompt = fixed_system_prompt_content;
// console.log("Using only fixed system prompt.");
} else if (hasSystemHistory) {
// Only overflow history exists (fixed system prompt was empty)
final_system_prompt = system_prompt_history_content;
// console.log("Using only history in system prompt slot.");
}
// If both are empty, final_system_prompt remains ""
// 6. Return the structured result
const result = {
system_prompt: final_system_prompt,
prompt: final_prompt
};
// console.log(`Final system_prompt length: ${result.system_prompt.length}`); // Debug log
// console.log(`Final prompt length: ${result.prompt.length}`); // Debug log
return result;
} catch (internalError) {
console.error("!!! CRITICAL ERROR inside convertMessagesToFalPrompt:", internalError);
console.error("!!! Failing messages input was:", JSON.stringify(messages, null, 2)); // Log the problematic input
// Re-throw the error to be caught by the main handler, indicating a setup failure
throw new Error(`Failed to process messages internally: ${internalError.message}`);
}
}
// === End convertMessagesToFalPrompt function ===
/**
* Makes a request to the Fal AI API, handling key rotation and retries on key-related errors.
* @param {object} falInput - The input object for the Fal AI API call.
* @param {boolean} [stream=false] - Whether to make a streaming request.
* @returns {Promise<object|AsyncIterable<object>>} The result object or async iterator for streams.
* @throws {Error} If the request fails after trying all valid keys, or if a non-key-related error occurs.
*/
async function makeFalRequestWithRetry(falInput, stream = false) {
let attempts = 0;
// Maximum attempts equals the initial number of keys
const maxAttempts = falKeys.length;
// Track keys attempted *within this specific request* to avoid infinite loops on retry logic issues
const attemptedKeysInThisRequest = new Set();
while (attempts < maxAttempts) {
const keyInfo = getNextValidKey(); // Get the next *valid* key info { key, index }
if (!keyInfo) {
// This should only happen if all keys are currently in the invalidKeys set
console.error("makeFalRequestWithRetry: No valid Fal AI keys remaining.");
throw new Error("No valid Fal AI keys available (all marked as invalid).");
}
// Prevent retrying the exact same key within this single request flow
if (attemptedKeysInThisRequest.has(keyInfo.key)) {
console.warn(`Key at index ${keyInfo.index} was already attempted for this request. Skipping to find next different key.`);
// Do not increment 'attempts' here as we didn't actually use the key. Let the loop find the next.
continue;
}
attemptedKeysInThisRequest.add(keyInfo.key);
attempts++; // Count this as a distinct attempt with a unique key for this request
try {
console.log(`Attempt ${attempts}/${maxAttempts}: Trying Fal Key index ${keyInfo.index}...`);
// *** CRITICAL: Reconfigure the global fal client with the selected key for THIS attempt ***
// WARNING: This reconfigures the GLOBAL client instance. In high-concurrency scenarios,
// this might lead to race conditions where one request configures the key just before another uses it.
// Consider instance pooling or check if fal-ai client offers per-request credentials if this becomes an issue.
console.warn(`Configuring GLOBAL fal client with key index ${keyInfo.index}. Review concurrency implications.`);
fal.config({ credentials: keyInfo.key }); // Use the specific key for this attempt
if (stream) {
// Use the now-configured global 'fal' object for the stream request
const falStream = await fal.stream("fal-ai/any-llm", { input: falInput });
console.log(`Successfully initiated stream with key index ${keyInfo.index}.`);
// Success! Return the stream iterator directly for the caller to handle
return falStream;
} else {
// Use the now-configured global 'fal' object for the non-stream request
console.log(`Executing non-stream request with key index ${keyInfo.index}...`);
// Use fal.subscribe (or appropriate non-stream method)
const result = await fal.subscribe("fal-ai/any-llm", {
input: falInput,
logs: true // Enable logs if helpful for debugging Fal side
});
console.log(`Successfully received non-stream result with key index ${keyInfo.index}.`);
// Optional: Check for errors *within* a successful-looking response structure
if (result && result.error) {
console.error(`Fal AI returned an error object within the non-stream result payload (Key Index ${keyInfo.index}):`, result.error);
// Decide if this specific payload error should also invalidate the key
if (isKeyRelatedError(result.error)) { // Reuse the checker
console.warn(`Marking Fal Key index ${keyInfo.index} as invalid due to error in response payload.`);
invalidKeys.add(keyInfo.key);
continue; // Go to the next iteration of the while loop (try next key)
} else {
// Throw an error that will be caught by the outer handler, not retried with other keys
throw new Error(`Fal AI error reported in result payload: ${JSON.stringify(result.error)}`);
}
}
// Success! Return the result
return result;
}
} catch (error) {
console.error(`Error caught using Fal Key index ${keyInfo.index}:`, error.message || error);
// Check if the caught error indicates the key is invalid
if (isKeyRelatedError(error)) {
console.warn(`Marking Fal Key index ${keyInfo.index} as invalid due to caught error.`);
// **ACTION: Add the failed key to the set of invalid keys**
invalidKeys.add(keyInfo.key);
// Continue to the next iteration of the while loop to try another key
} else {
// Error does not appear key-related (e.g., network issue, bad input format, internal Fal server error)
// Do not retry with other keys for this type of error. Fail the request immediately.
console.error("Error does not appear to be key-related. Failing request without further key retries.");
throw error; // Re-throw the original error to be caught by the main endpoint handler
}
}
} // End while loop
// If the loop finishes without returning/throwing earlier, it means all available keys were tried and failed with key-related errors
throw new Error(`Request failed after trying ${attempts} unique Fal key(s). All failed with key-related errors or were already marked invalid.`);
}
// POST /v1/chat/completions endpoint - Handles chat requests, uses key rotation/failover
app.post('/v1/chat/completions', async (req, res) => {
// Extract parameters from request body
const { model, messages, stream = false, reasoning = false, ...restOpenAIParams } = req.body; // restOpenAIParams currently ignored but captured
console.log(`--> POST /v1/chat/completions | Model: ${model} | Stream: ${stream}`);
// Basic Input Validation
if (!FAL_SUPPORTED_MODELS.includes(model)) {
// Log warning but allow attempt if model isn't in the known list
console.warn(`Warning: Requested model '${model}' is not in the explicitly supported list. Proxy will still attempt the request.`);
}
if (!model || !messages || !Array.isArray(messages) || messages.length === 0) {
console.error("Invalid request: Missing 'model' or 'messages' array is empty/invalid.");
return res.status(400).json({ error: 'Bad Request: `model` and a non-empty `messages` array are required.' });
}
try {
// --- Prepare Fal AI Input using the conversion function ---
// This step might throw an error if messages are invalid, caught by the outer catch block
const { prompt, system_prompt } = convertMessagesToFalPrompt(messages);
const falInput = {
model: model, // Pass the requested model
prompt: prompt, // The main prompt constructed from recent history
// Only include system_prompt if it has content
...(system_prompt && system_prompt.length > 0 && { system_prompt: system_prompt }),
reasoning: !!reasoning, // Ensure boolean, pass reasoning flag if provided
};
// console.debug("Prepared Fal Input:", JSON.stringify(falInput, null, 2)); // Verbose debug log
console.log("Attempting Fal request with key rotation/retry logic...");
console.log(`Prepared Input Lengths - System Prompt: ${system_prompt?.length || 0}, Prompt: ${prompt?.length || 0}`);
// --- Handle Stream vs Non-Stream using the retry helper function ---
if (stream) {
// Set headers for Server-Sent Events (SSE)
res.setHeader('Content-Type', 'text/event-stream; charset=utf-8');
res.setHeader('Cache-Control', 'no-cache');
res.setHeader('Connection', 'keep-alive');
res.setHeader('Access-Control-Allow-Origin', '*'); // Adjust CORS for production if needed
res.flushHeaders(); // Send headers immediately
let previousOutput = ''; // Track previous output for delta calculation
let falStream; // Variable to hold the stream iterator
try {
// **Initiate the stream using the retry helper**
falStream = await makeFalRequestWithRetry(falInput, true);
// Process the stream events asynchronously
for await (const event of falStream) {
// Safely extract data from the event
const currentOutput = (event && typeof event.output === 'string') ? event.output : '';
// Default to partial=true if missing
const isPartial = (event && typeof event.partial === 'boolean') ? event.partial : true;
const errorInfo = (event && event.error) ? event.error : null; // Check for errors within the stream event
// Handle errors reported *within* a stream event payload
if (errorInfo) {
console.error("Error received *within* fal stream event payload:", errorInfo);
// Optionally send an error chunk to the client
const errorChunk = {
id: `chatcmpl-${Date.now()}-error`, object: "chat.completion.chunk", created: Math.floor(Date.now() / 1000), model: model,
choices: [{ index: 0, delta: {}, finish_reason: "error", message: { role: 'assistant', content: `Fal Stream Event Error: ${JSON.stringify(errorInfo)}` } }]
};
// Check if stream is still writable before sending
if (!res.writableEnded) {
res.write(`data: ${JSON.stringify(errorChunk)}\n\n`);
} else {
console.warn("Stream already ended when trying to write stream event error.");
}
// Depending on the error, you might want to break or continue
// break; // Uncomment to stop processing on first stream error
}
// Calculate the delta (new content) since the last event
let deltaContent = '';
if (currentOutput.startsWith(previousOutput)) {
// Normal case: current output contains previous plus new content
deltaContent = currentOutput.substring(previousOutput.length);
} else if (currentOutput.length > 0) {
// Output mismatch or reset: send the entire current output as delta
console.warn("Fal stream output mismatch or reset detected. Sending full current output as delta.");
deltaContent = currentOutput;
previousOutput = ''; // Reset comparison base on mismatch
} // If currentOutput is empty, deltaContent remains empty
previousOutput = currentOutput; // Update for the next iteration
// Send OpenAI-compatible SSE chunk if there's new content or if it's the final chunk
if (deltaContent || !isPartial) {
const openAIChunk = {
id: `chatcmpl-${Date.now()}`, // Consider more robust ID generation if needed
object: "chat.completion.chunk",
created: Math.floor(Date.now() / 1000),
model: model, // Echo back the requested model
choices: [{
index: 0,
delta: { content: deltaContent }, // The new part of the content
// Set finish_reason only on the final chunk
finish_reason: isPartial === false ? "stop" : null
}]
// system_fingerprint is not provided by Fal, so omit or set to null
};
// Check if stream is still writable before sending
if (!res.writableEnded) {
res.write(`data: ${JSON.stringify(openAIChunk)}\n\n`);
} else {
console.warn("Stream already ended when trying to write data chunk.");
}
}
} // End for-await loop over falStream
// Send the final [DONE] marker to indicate stream completion
if (!res.writableEnded) {
res.write(`data: [DONE]\n\n`);
res.end(); // Close the connection
console.log("<-- Stream finished successfully and [DONE] sent.");
} else {
console.log("<-- Stream processing finished, but connection was already ended before [DONE].");
}
} catch (streamError) {
// Catches errors from makeFalRequestWithRetry OR the stream iteration itself (e.g., network drop)
console.error('Error during stream request processing:', streamError.message || streamError);
try {
// If headers haven't been sent, the error likely happened during initial connection (makeFalRequestWithRetry)
if (!res.headersSent) {
const errorMessage = (streamError instanceof Error) ? streamError.message : JSON.stringify(streamError);
// Use 502 Bad Gateway for upstream failures (like all keys failing)
res.status(502).json({
error: 'Failed to initiate Fal stream',
details: errorMessage // Include the error message from the helper
});
console.log("<-- Stream initiation failed response sent (502).");
} else if (!res.writableEnded) {
// Stream started but failed mid-way. Try to send an error message within the stream context.
const errorDetails = (streamError instanceof Error) ? streamError.message : JSON.stringify(streamError);
// Send an error object in the SSE stream format
res.write(`data: ${JSON.stringify({ error: { message: "Stream processing error after initiation", type: "proxy_error", details: errorDetails } })}\n\n`);
res.write(`data: [DONE]\n\n`); // Still send DONE after error for robust client handling
res.end();
console.log("<-- Stream error sent within stream, stream ended.");
} else {
// Stream already ended, just log the error server-side.
console.log("<-- Stream error occurred, but connection was already ended.");
}
} catch (finalError) {
// Error trying to send the error message itself (rare)
console.error('Error sending stream error message to client:', finalError);
// Ensure response is ended if possible
if (!res.writableEnded) { res.end(); }
}
}
} else {
// --- Non-Stream Request ---
try {
// **Get the result using the retry helper**
const result = await makeFalRequestWithRetry(falInput, false);
// console.debug("Received non-stream result via retry function:", JSON.stringify(result, null, 2)); // Verbose debug
// --- Construct OpenAI compatible response ---
const openAIResponse = {
id: `chatcmpl-${result.requestId || Date.now()}`, // Use Fal's requestId if available
object: "chat.completion",
created: Math.floor(Date.now() / 1000),
model: model, // Echo back the requested model
choices: [{
index: 0,
message: {
role: "assistant",
content: result.output || "" // Ensure content is a string, default to empty if missing
},
finish_reason: "stop" // Assume 'stop' for successful non-stream completion
}],
usage: { // Fal doesn't provide token usage, return nulls
prompt_tokens: null,
completion_tokens: null,
total_tokens: null
},
system_fingerprint: null, // Fal doesn't provide this
// Include Fal specific reasoning if present and requested
...(result.reasoning && { fal_reasoning: result.reasoning }),
};
res.json(openAIResponse);
console.log("<-- Non-stream response sent successfully.");
} catch (error) {
// Catches errors from makeFalRequestWithRetry (e.g., all keys failed or a non-key-related Fal error)
console.error('Error during non-stream request processing:', error.message || error);
// Check if response can still be sent
if (!res.headersSent) {
const errorMessage = (error instanceof Error) ? error.message : JSON.stringify(error);
// Customize error message if it's the specific "all keys failed" error
const finalMessage = errorMessage.includes("No valid Fal AI keys available") || errorMessage.includes("Request failed after trying")
? `Fal request failed: ${errorMessage}` // More direct message
: `Internal Server Error processing Fal request: ${errorMessage}`;
// Use 502 Bad Gateway to indicate upstream failure
res.status(502).json({ error: 'Fal Request Failed', details: finalMessage });
console.log("<-- Non-stream error response sent (502).");
} else {
// Should be rare for non-stream, but log if headers were already sent
console.error("Headers already sent for non-stream error response? This is unexpected.");
if (!res.writableEnded) { res.end(); } // Attempt to end response if possible
}
}
}
} catch (error) {
// Catch errors occurring *before* the Fal request attempt
// (e.g., error during `convertMessagesToFalPrompt`, JSON parsing errors)
console.error('Unhandled error before initiating Fal request (likely setup or input conversion):', error.message || error);
if (!res.headersSent) {
const errorMessage = (error instanceof Error) ? error.message : JSON.stringify(error);
// Use 500 Internal Server Error for issues within the proxy itself
res.status(500).json({ error: 'Internal Server Error in Proxy Setup', details: errorMessage });
console.log("<-- Proxy setup error response sent (500).");
} else {
console.error("Headers already sent when catching setup error. Ending response.");
if (!res.writableEnded) { res.end(); }
}
}
});
// Start the Express server
app.listen(PORT, () => {
console.log(`=====================================================================`);
console.log(` Fal OpenAI Proxy Server (Multi-Key Rotation & Failover)`);
console.log(`---------------------------------------------------------------------`);
console.log(` Listening on port : ${PORT}`);
console.log(` Reading Fal Keys from : FAL_KEY environment variable (comma-separated)`);
console.log(` Loaded Keys Count : ${falKeys.length}`);
console.log(` Invalid Keys Set : Initialized (size: ${invalidKeys.size})`);
console.log(` Proxy API Key Auth : ${API_KEY ? 'Enabled (using API_KEY env var)' : 'DISABLED'}`);
console.log(` Input Limits : System Prompt=${SYSTEM_PROMPT_LIMIT}, Prompt=${PROMPT_LIMIT}`);
console.log(` Concurrency Warning : Global Fal client reconfigured per request attempt!`);
console.log(`---------------------------------------------------------------------`);
console.log(` Endpoints Available:`);
console.log(` POST http://localhost:${PORT}/v1/chat/completions`);
console.log(` GET http://localhost:${PORT}/v1/models`);
console.log(`=====================================================================`);
});
// Root path handler for basic health check / info
app.get('/', (req, res) => {
res.send(`Fal OpenAI Proxy (Multi-Key Rotation/Failover from FAL_KEY) is running. Loaded ${falKeys.length} key(s). Currently ${invalidKeys.size} key(s) marked as invalid.`);
}); |