Files
open-gsio/workers/site/services/ChatService.ts
geoffsee 33679583af init
2025-05-22 23:14:01 -04:00

519 lines
17 KiB
TypeScript

import { flow, getSnapshot, types } from "mobx-state-tree";
import OpenAI from "openai";
import ChatSdk from "../sdk/chat-sdk";
import Message from "../models/Message";
import O1Message from "../models/O1Message";
import {
getModelFamily,
ModelFamily,
} from "../../../src/components/chat/SupportedModels";
import { OpenAiChatSdk } from "../sdk/models/openai";
import { GroqChatSdk } from "../sdk/models/groq";
import { ClaudeChatSdk } from "../sdk/models/claude";
import { FireworksAiChatSdk } from "../sdk/models/fireworks";
import handleStreamData from "../sdk/handleStreamData";
import { GoogleChatSdk } from "../sdk/models/google";
import { XaiChatSdk } from "../sdk/models/xai";
import { CerebrasSdk } from "../sdk/models/cerebras";
import { CloudflareAISdk } from "../sdk/models/cloudflareAi";
export interface StreamParams {
env: Env;
openai: OpenAI;
messages: any[];
model: string;
systemPrompt: string;
preprocessedContext: any;
attachments: any[];
tools: any[];
disableWebhookGeneration: boolean;
maxTokens: number;
}
interface StreamHandlerParams {
controller: ReadableStreamDefaultController;
encoder: TextEncoder;
webhook?: { url: string; payload: unknown };
dynamicContext?: any;
}
const ChatService = types
.model("ChatService", {
openAIApiKey: types.optional(types.string, ""),
openAIBaseURL: types.optional(types.string, ""),
activeStreams: types.optional(
types.map(
types.model({
name: types.optional(types.string, ""),
maxTokens: types.optional(types.number, 0),
systemPrompt: types.optional(types.string, ""),
model: types.optional(types.string, ""),
messages: types.optional(types.array(types.frozen()), []),
attachments: types.optional(types.array(types.frozen()), []),
tools: types.optional(types.array(types.frozen()), []),
disableWebhookGeneration: types.optional(types.boolean, false),
}),
),
),
maxTokens: types.number,
systemPrompt: types.string,
})
.volatile((self) => ({
openai: {} as OpenAI,
env: {} as Env,
webhookStreamActive: false,
}))
.actions((self) => {
const createMessageInstance = (message: any) => {
if (typeof message.content === "string") {
return Message.create({
role: message.role,
content: message.content,
});
}
if (Array.isArray(message.content)) {
const m = O1Message.create({
role: message.role,
content: message.content.map((item) => ({
type: item.type,
text: item.text,
})),
});
return m;
}
throw new Error("Unsupported message format");
};
const handleWebhookProcessing = async ({
controller,
encoder,
webhook,
dynamicContext,
}: StreamHandlerParams) => {
console.log("handleWebhookProcessing::start");
if (!webhook) return;
console.log("handleWebhookProcessing::[Loading Live Search]");
dynamicContext.append("\n## Live Search\n~~~markdown\n");
for await (const chunk of self.streamWebhookData({ webhook })) {
controller.enqueue(encoder.encode(chunk));
dynamicContext.append(chunk);
}
dynamicContext.append("\n~~~\n");
console.log(
`handleWebhookProcessing::[Finished loading Live Search!][length: ${dynamicContext.content.length}]`,
);
ChatSdk.sendDoubleNewline(controller, encoder);
console.log("handleWebhookProcessing::exit");
};
const createStreamParams = async (
streamConfig: any,
dynamicContext: any,
durableObject: any,
): Promise<StreamParams> => {
return {
env: self.env,
openai: self.openai,
messages: streamConfig.messages.map(createMessageInstance),
model: streamConfig.model,
systemPrompt: streamConfig.systemPrompt,
preprocessedContext: getSnapshot(dynamicContext),
attachments: streamConfig.attachments ?? [],
tools: streamConfig.tools ?? [],
disableWebhookGeneration: true,
maxTokens: await durableObject.dynamicMaxTokens(
streamConfig.messages,
2000,
),
};
};
const modelHandlers = {
openai: (params: StreamParams, dataHandler: Function) =>
OpenAiChatSdk.handleOpenAiStream(params, dataHandler),
groq: (params: StreamParams, dataHandler: Function) =>
GroqChatSdk.handleGroqStream(params, dataHandler),
claude: (params: StreamParams, dataHandler: Function) =>
ClaudeChatSdk.handleClaudeStream(params, dataHandler),
fireworks: (params: StreamParams, dataHandler: Function) =>
FireworksAiChatSdk.handleFireworksStream(params, dataHandler),
google: (params: StreamParams, dataHandler: Function) =>
GoogleChatSdk.handleGoogleStream(params, dataHandler),
xai: (params: StreamParams, dataHandler: Function) =>
XaiChatSdk.handleXaiStream(params, dataHandler),
cerebras: (params: StreamParams, dataHandler: Function) =>
CerebrasSdk.handleCerebrasStream(params, dataHandler),
cloudflareAI: (params: StreamParams, dataHandler: Function) =>
CloudflareAISdk.handleCloudflareAIStream(params, dataHandler),
};
return {
setActiveStream(streamId: string, stream: any) {
const validStream = {
name: stream?.name || "Unnamed Stream",
maxTokens: stream?.maxTokens || 0,
systemPrompt: stream?.systemPrompt || "",
model: stream?.model || "",
messages: stream?.messages || [],
attachments: stream?.attachments || [],
tools: stream?.tools || [],
disableWebhookGeneration: stream?.disableWebhookGeneration || false,
};
self.activeStreams.set(streamId, validStream);
},
removeActiveStream(streamId: string) {
self.activeStreams.delete(streamId);
},
setEnv(env: Env) {
self.env = env;
self.openai = new OpenAI({
apiKey: self.openAIApiKey,
baseURL: self.openAIBaseURL,
});
},
handleChatRequest: async (request: Request) => {
return ChatSdk.handleChatRequest(request, {
openai: self.openai,
env: self.env,
systemPrompt: self.systemPrompt,
maxTokens: self.maxTokens,
});
},
setWebhookStreamActive(value) {
self.webhookStreamActive = value;
},
streamWebhookData: async function* ({ webhook }) {
console.log("streamWebhookData::start");
if (self.webhookStreamActive) {
return;
}
const queue: string[] = [];
let resolveQueueItem: Function;
let finished = false;
let errorOccurred: Error | null = null;
const dataPromise = () =>
new Promise<void>((resolve) => {
resolveQueueItem = resolve;
});
let currentPromise = dataPromise();
const eventSource = new EventSource(webhook.url.trim());
console.log("streamWebhookData::setWebhookStreamActive::true");
self.setWebhookStreamActive(true);
try {
ChatSdk.handleWebhookStream(eventSource, (data) => {
const formattedData = `data: ${JSON.stringify(data)}\n\n`;
queue.push(formattedData);
if (resolveQueueItem) resolveQueueItem();
currentPromise = dataPromise();
})
.then(() => {
finished = true;
if (resolveQueueItem) resolveQueueItem();
})
.catch((err) => {
console.log(
`chatService::streamWebhookData::STREAM_ERROR::${err}`,
);
errorOccurred = err;
if (resolveQueueItem) resolveQueueItem();
});
while (!finished || queue.length > 0) {
if (queue.length > 0) {
yield queue.shift()!;
} else if (errorOccurred) {
throw errorOccurred;
} else {
await currentPromise;
}
}
self.setWebhookStreamActive(false);
eventSource.close();
console.log(`chatService::streamWebhookData::complete`);
} catch (error) {
console.log(`chatService::streamWebhookData::error`);
eventSource.close();
self.setWebhookStreamActive(false);
console.error("Error while streaming webhook data:", error);
throw error;
}
},
/**
* runModelHandler
* Selects the correct model handler and invokes it.
*/
async runModelHandler(params: {
streamConfig: any;
streamParams: any;
controller: ReadableStreamDefaultController;
encoder: TextEncoder;
streamId: string;
}) {
const { streamConfig, streamParams, controller, encoder, streamId } =
params;
const modelFamily = getModelFamily(streamConfig.model);
console.log(
`chatService::handleSseStream::ReadableStream::modelFamily::${modelFamily}`,
);
const handler = modelHandlers[modelFamily as ModelFamily];
if (handler) {
try {
console.log(
`chatService::handleSseStream::ReadableStream::${streamId}::handler::start`,
);
await handler(streamParams, handleStreamData(controller, encoder));
console.log(
`chatService::handleSseStream::ReadableStream::${streamId}::handler::finish`,
);
} catch (error) {
const message = error.message.toLowerCase();
if (
message.includes("413 ") ||
message.includes("maximum") ||
message.includes("too long") ||
message.includes("too large")
) {
throw new ClientError(
`Error! Content length exceeds limits. Try shortening your message, removing any attached files, or editing an earlier message instead.`,
413,
{
model: streamConfig.model,
maxTokens: streamParams.maxTokens,
},
);
}
if (message.includes("429 ")) {
throw new ClientError(
`Error! Rate limit exceeded. Wait a few minutes before trying again.`,
429,
{
model: streamConfig.model,
maxTokens: streamParams.maxTokens,
},
);
}
if (message.includes("404")) {
throw new ClientError(
`Something went wrong, try again.`,
413,
{},
);
}
throw error;
/*
'413 Request too large for model `mixtral-8x7b-32768` in organization `org_01htjxws48fm0rbbg5gnkgmbrh` service tier `on_demand` on tokens per minute (TPM): Limit 5000, Requested 49590, please reduce your message size and try again. Visit https://console.groq.com/docs/rate-limits for more information.'
*/
}
}
},
/**
* handleWebhookIfNeeded
* Checks if a webhook exists, and if so, processes it.
*/
async handleWebhookIfNeeded(params: {
savedStreamConfig: string;
controller: ReadableStreamDefaultController;
encoder: TextEncoder;
}) {
const { savedStreamConfig, controller, encoder, dynamicContext } =
params;
const config = JSON.parse(savedStreamConfig);
const webhook = config?.webhooks?.[0];
if (webhook) {
console.log(
`chatService::handleSseStream::ReadableStream::webhook:start`,
);
await handleWebhookProcessing({
controller,
encoder,
webhook,
dynamicContext,
});
console.log(
`chatService::handleSseStream::ReadableStream::webhook::end`,
);
}
},
createSseReadableStream(params: {
streamId: string;
streamConfig: any;
savedStreamConfig: string;
durableObject: any;
}) {
const { streamId, streamConfig, savedStreamConfig, durableObject } =
params;
return new ReadableStream({
async start(controller) {
console.log(
`chatService::handleSseStream::ReadableStream::${streamId}::open`,
);
const encoder = new TextEncoder();
try {
const dynamicContext = Message.create(
streamConfig.preprocessedContext,
);
await self.handleWebhookIfNeeded({
savedStreamConfig,
controller,
encoder,
dynamicContext: dynamicContext,
});
const streamParams = await createStreamParams(
streamConfig,
dynamicContext,
durableObject,
);
try {
await self.runModelHandler({
streamConfig,
streamParams,
controller,
encoder,
streamId,
});
} catch (e) {
console.log("error caught at runModelHandler");
throw e;
}
} catch (error) {
console.error(
`chatService::handleSseStream::${streamId}::Error`,
error,
);
if (error instanceof ClientError) {
controller.enqueue(
encoder.encode(
`data: ${JSON.stringify({ type: "error", error: error.message })}\n\n`,
),
);
} else {
controller.enqueue(
encoder.encode(
`data: ${JSON.stringify({ type: "error", error: "Server error" })}\n\n`,
),
);
}
controller.close();
} finally {
try {
controller.close();
} catch (_) {}
}
},
});
},
handleSseStream: flow(function* (
streamId: string,
): Generator<Promise<string>, Response, unknown> {
console.log(`chatService::handleSseStream::enter::${streamId}`);
if (self.activeStreams.has(streamId)) {
console.log(
`chatService::handleSseStream::${streamId}::[stream already active]`,
);
return new Response("Stream already active", { status: 409 });
}
const objectId = self.env.SITE_COORDINATOR.idFromName("stream-index");
const durableObject = self.env.SITE_COORDINATOR.get(objectId);
const savedStreamConfig = yield durableObject.getStreamData(streamId);
if (!savedStreamConfig) {
return new Response("Stream not found", { status: 404 });
}
const streamConfig = JSON.parse(savedStreamConfig);
console.log(
`chatService::handleSseStream::${streamId}::[stream configured]`,
);
const stream = self.createSseReadableStream({
streamId,
streamConfig,
savedStreamConfig,
durableObject,
});
const [processingStream, responseStream] = stream.tee();
self.setActiveStream(streamId, {});
processingStream.pipeTo(
new WritableStream({
close() {
console.log(
`chatService::handleSseStream::${streamId}::[stream closed]`,
);
},
}),
);
return new Response(responseStream, {
headers: {
"Content-Type": "text/event-stream",
"Cache-Control": "no-cache",
Connection: "keep-alive",
},
});
}),
};
});
/**
* ClientError
* A custom construct for sending client-friendly errors via the controller in a structured and controlled manner.
*/
export class ClientError extends Error {
public statusCode: number;
public details: Record<string, any>;
constructor(
message: string,
statusCode: number,
details: Record<string, any> = {},
) {
super(message);
this.name = "ClientError";
this.statusCode = statusCode;
this.details = details;
Object.setPrototypeOf(this, ClientError.prototype);
}
/**
* Formats the error for SSE-compatible data transmission.
*/
public formatForSSE(): string {
return JSON.stringify({
type: "error",
message: this.message,
details: this.details,
statusCode: this.statusCode,
});
}
}
export default ChatService;