import {flow, getSnapshot, types} from 'mobx-state-tree'; import OpenAI from 'openai'; import ChatSdk from '../sdk/chat-sdk'; import Message from "../models/Message"; import O1Message from "../models/O1Message"; import {getModelFamily, ModelFamily} from "../../../src/components/chat/SupportedModels"; import {OpenAiChatSdk} from "../sdk/models/openai"; import {GroqChatSdk} from "../sdk/models/groq"; import {ClaudeChatSdk} from "../sdk/models/claude"; import {FireworksAiChatSdk} from "../sdk/models/fireworks"; import handleStreamData from "../sdk/handleStreamData"; import {GoogleChatSdk} from "../sdk/models/google"; import {XaiChatSdk} from "../sdk/models/xai"; import {CerebrasSdk} from "../sdk/models/cerebras"; import {CloudflareAISdk} from "../sdk/models/cloudflareAi"; export interface StreamParams { env: Env; openai: OpenAI; messages: any[]; model: string; systemPrompt: string; preprocessedContext: any; attachments: any[]; tools: any[]; disableWebhookGeneration: boolean; maxTokens: number; } interface StreamHandlerParams { controller: ReadableStreamDefaultController; encoder: TextEncoder; webhook?: { url: string, payload: unknown }; dynamicContext?: any; } const activeStreamType = types.model({ name: types.optional(types.string, ""), maxTokens: types.optional(types.number, 0), systemPrompt: types.optional(types.string, ""), model: types.optional(types.string, ""), messages: types.optional(types.array(types.frozen()), []), attachments: types.optional(types.array(types.frozen()), []), tools: types.optional(types.array(types.frozen()), []), disableWebhookGeneration: types.optional(types.boolean, false) }); const activeStreamsMap = types.map( activeStreamType, ); const ChatService = types .model('ChatService', { openAIApiKey: types.optional(types.string, ""), openAIBaseURL: types.optional(types.string, ""), activeStreams: types.optional( activeStreamsMap, {} // Correct initialization ), maxTokens: types.number, systemPrompt: types.string }) .volatile(self => ({ openai: {} as OpenAI, env: {} as Env, webhookStreamActive: false })) .actions(self => { // Helper functions const createMessageInstance = (message: any) => { if (typeof message.content === 'string') { return Message.create({ role: message.role, content: message.content, }); } if (Array.isArray(message.content)) { const m = O1Message.create({ role: message.role, content: message.content.map(item => ({ type: item.type, text: item.text })), }); return m; } throw new Error('Unsupported message format'); }; const handleAgentProcess = async ( {controller, encoder, webhook, dynamicContext}: StreamHandlerParams ) => { if (!webhook) return; dynamicContext.append("\n## Agent Results\n~~~markdown\n"); for await (const chunk of self.streamAgentData({webhook})) { controller.enqueue(encoder.encode(chunk)); dynamicContext.append(chunk); } dynamicContext.append("\n~~~\n"); ChatSdk.sendDoubleNewline(controller, encoder); }; const createStreamParams = async ( streamConfig: any, dynamicContext: any, durableObject: any ): Promise => { return { env: self.env, openai: self.openai, messages: streamConfig.messages.map(createMessageInstance), model: streamConfig.model, systemPrompt: streamConfig.systemPrompt, preprocessedContext: getSnapshot(dynamicContext), attachments: streamConfig.attachments ?? [], tools: streamConfig.tools ?? [], disableWebhookGeneration: true, maxTokens: await durableObject.dynamicMaxTokens( streamConfig.messages, 2000 ), } }; const modelHandlers = { openai: (params: StreamParams, dataHandler: Function) => OpenAiChatSdk.handleOpenAiStream(params, dataHandler), groq: (params: StreamParams, dataHandler: Function) => GroqChatSdk.handleGroqStream(params, dataHandler), claude: (params: StreamParams, dataHandler: Function) => ClaudeChatSdk.handleClaudeStream(params, dataHandler), fireworks: (params: StreamParams, dataHandler: Function) => FireworksAiChatSdk.handleFireworksStream(params, dataHandler), google: (params: StreamParams, dataHandler: Function) => GoogleChatSdk.handleGoogleStream(params, dataHandler), xai: (params: StreamParams, dataHandler: Function) => XaiChatSdk.handleXaiStream(params, dataHandler), cerebras: (params: StreamParams, dataHandler: Function) => CerebrasSdk.handleCerebrasStream(params, dataHandler), cloudflareAI: (params: StreamParams, dataHandler: Function) => CloudflareAISdk.handleCloudflareAIStream(params, dataHandler) }; return { setActiveStream(streamId: string, stream: any) { const validStream = { name: stream?.name || "Unnamed Stream", maxTokens: stream?.maxTokens || 0, systemPrompt: stream?.systemPrompt || "", model: stream?.model || "", messages: stream?.messages || [], attachments: stream?.attachments || [], tools: stream?.tools || [], disableWebhookGeneration: stream?.disableWebhookGeneration || false, }; self.activeStreams.set(streamId, validStream); }, removeActiveStream(streamId: string) { self.activeStreams.delete(streamId); }, setEnv(env: Env) { self.env = env; self.openai = new OpenAI({ apiKey: self.openAIApiKey, baseURL: self.openAIBaseURL, }); }, handleChatRequest: async (request: Request) => { return ChatSdk.handleChatRequest(request, { openai: self.openai, env: self.env, systemPrompt: self.systemPrompt, maxTokens: self.maxTokens }); }, setWebhookStreamActive(value) { self.webhookStreamActive = value; }, streamAgentData: async function* ({webhook}) { console.log("streamAgentData::start"); if (self.webhookStreamActive) { return } const queue: string[] = []; let resolveQueueItem: Function; let finished = false; let errorOccurred: Error | null = null; const dataPromise = () => new Promise((resolve) => { resolveQueueItem = resolve; }); let currentPromise = dataPromise(); const eventSource = new EventSource(webhook.url.trim()); console.log("streamAgentData::setWebhookStreamActive::true"); self.setWebhookStreamActive(true) try { ChatSdk.handleAgentStream(eventSource, (data) => { const formattedData = `data: ${JSON.stringify(data)}\n\n`; queue.push(formattedData); if (resolveQueueItem) resolveQueueItem(); currentPromise = dataPromise(); }).then(() => { finished = true; if (resolveQueueItem) resolveQueueItem(); }).catch((err) => { console.log(`chatService::streamAgentData::STREAM_ERROR::${err}`); errorOccurred = err; if (resolveQueueItem) resolveQueueItem(); }); while (!finished || queue.length > 0) { if (queue.length > 0) { yield queue.shift()!; } else if (errorOccurred) { throw errorOccurred; } else { await currentPromise; } } self.setWebhookStreamActive(false); eventSource.close(); // console.log(`chatService::streamAgentData::complete`); } catch (error) { console.log(`chatService::streamAgentData::error`); eventSource.close(); self.setWebhookStreamActive(false); console.error("Error while streaming webhook data:", error); throw error; } }, /** * runModelHandler * Selects the correct model handler and invokes it. */ async runModelHandler(params: { streamConfig: any; streamParams: any; controller: ReadableStreamDefaultController; encoder: TextEncoder; streamId: string; }) { const {streamConfig, streamParams, controller, encoder, streamId} = params; const modelFamily = getModelFamily(streamConfig.model); console.log( `chatService::handleSseStream::ReadableStream::modelFamily::${modelFamily}` ); const handler = modelHandlers[modelFamily as ModelFamily]; if (handler) { try { console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::start`); await handler(streamParams, handleStreamData(controller, encoder)); console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::finish`); } catch (error) { const message = error.message.toLowerCase(); if(message.includes("413 ") || (message.includes("maximum") || message.includes("too long") || message.includes("too large") )) { throw new ClientError(`Error! Content length exceeds limits. Try shortening your message, removing any attached files, or editing an earlier message instead.`, 413, {model: streamConfig.model, maxTokens: streamParams.maxTokens}) // throw new Error(`Max tokens exceeded for model ${streamConfig.model}`) } if(message.includes("429 ")) { throw new ClientError(`Error! Rate limit exceeded. Wait a few minutes before trying again.`, 429, {model: streamConfig.model, maxTokens: streamParams.maxTokens}) // throw new Error(`Max tokens exceeded for model ${streamConfig.model}`) } if (message.includes("404")) { throw new ClientError(`Something went wrong, try again.`, 413, {}) // throw new Error(`Max tokens exceeded for model ${streamConfig.model}`) } throw error; /* '413 Request too large for model `mixtral-8x7b-32768` in organization `org_01htjxws48fm0rbbg5gnkgmbrh` service tier `on_demand` on tokens per minute (TPM): Limit 5000, Requested 49590, please reduce your message size and try again. Visit https://console.groq.com/docs/rate-limits for more information.' */ } } }, /** * bootstrapAgents * Checks if an agent exists, and if so, bootstraps it. */ async bootstrapAgents(params: { savedStreamConfig: string; controller: ReadableStreamDefaultController; encoder: TextEncoder; dynamicContext: any; // or more specific type }) { const {savedStreamConfig, controller, encoder, dynamicContext} = params; const config = JSON.parse(savedStreamConfig); const webhook = config?.webhooks?.[0]; if (webhook) { console.log(`chatService::handleSseStream::ReadableStream::webhook:start`); await handleAgentProcess({ controller, encoder, webhook, dynamicContext, }); console.log(`chatService::handleSseStream::ReadableStream::webhook::end`); } }, createSseReadableStream(params: { streamId: string; streamConfig: any; savedStreamConfig: string; durableObject: any; }) { const { streamId, streamConfig, savedStreamConfig, durableObject } = params; return new ReadableStream({ async start(controller) { console.log(`chatService::handleSseStream::ReadableStream::${streamId}::open`); const encoder = new TextEncoder(); try { const dynamicContext = Message.create(streamConfig.preprocessedContext); // Process agents if configured await self.bootstrapAgents({ savedStreamConfig, controller, encoder, dynamicContext: dynamicContext, }); // Process the stream data using the appropriate handler const streamParams = await createStreamParams( streamConfig, dynamicContext, durableObject ); try { await self.runModelHandler({ streamConfig, streamParams, controller, encoder, streamId, }); } catch (e) { console.log("error caught at runModelHandler") throw e; } } catch (error) { console.error(`chatService::handleSseStream::${streamId}::Error`, error); if(error instanceof ClientError) { controller.enqueue( encoder.encode(`data: ${JSON.stringify({ type: 'error', error: error.message })}\n\n`) ); } else { controller.enqueue( encoder.encode(`data: ${JSON.stringify({ type: 'error', error: "Server error" })}\n\n`) ); } controller.close(); } finally { try { controller.close(); } catch (_) {} } }, }); }, handleSseStream: flow(function* (streamId: string): Generator, Response, unknown> { console.log(`chatService::handleSseStream::enter::${streamId}`); // Check if a stream is already active for this ID if (self.activeStreams.has(streamId)) { console.log(`chatService::handleSseStream::${streamId}::[stream already active]`); return new Response('Stream already active', { status: 409 }); } // Retrieve the stream configuration from the durable object const objectId = self.env.SITE_COORDINATOR.idFromName('stream-index'); const durableObject = self.env.SITE_COORDINATOR.get(objectId); const savedStreamConfig = yield durableObject.getStreamData(streamId); if (!savedStreamConfig) { return new Response('Stream not found', { status: 404 }); } const streamConfig = JSON.parse(savedStreamConfig); console.log(`chatService::handleSseStream::${streamId}::[stream configured]`); const stream = self.createSseReadableStream({ streamId, streamConfig, savedStreamConfig, durableObject, }); // Use `tee()` to create two streams: one for processing and one for the response const [processingStream, responseStream] = stream.tee(); self.setActiveStream(streamId, { ...streamConfig, }); processingStream.pipeTo( new WritableStream({ close() { self.removeActiveStream(streamId); }, }) ); // Return the second stream as the response return new Response(responseStream, { headers: { 'Content-Type': 'text/event-stream', 'Cache-Control': 'no-cache', 'Connection': 'keep-alive', }, }); }), }; }); /** * ClientError * A custom construct for sending client-friendly errors via the controller in a structured and controlled manner. */ export class ClientError extends Error { public statusCode: number; public details: Record; constructor(message: string, statusCode: number, details: Record = {}) { super(message); this.name = 'ClientError'; this.statusCode = statusCode; this.details = details; Object.setPrototypeOf(this, ClientError.prototype); } /** * Formats the error for SSE-compatible data transmission. */ public formatForSSE(): string { return JSON.stringify({ type: 'error', message: this.message, details: this.details, statusCode: this.statusCode, }); } } export default ChatService;