mirror of
https://github.com/geoffsee/open-gsio.git
synced 2025-09-08 22:56:46 +00:00

This commit removes the preprocessing workflow, its operators, intent service, and associated functionality. Additionally, redundant logging and unnecessary comments have been cleaned up in the ChatService for better readability and maintainability.
469 lines
17 KiB
TypeScript
469 lines
17 KiB
TypeScript
import {flow, getSnapshot, types} from 'mobx-state-tree';
|
|
import OpenAI from 'openai';
|
|
import ChatSdk from '../sdk/chat-sdk';
|
|
import Message from "../models/Message";
|
|
import O1Message from "../models/O1Message";
|
|
import {getModelFamily, ModelFamily} from "../../../src/components/chat/SupportedModels";
|
|
import {OpenAiChatSdk} from "../sdk/models/openai";
|
|
import {GroqChatSdk} from "../sdk/models/groq";
|
|
import {ClaudeChatSdk} from "../sdk/models/claude";
|
|
import {FireworksAiChatSdk} from "../sdk/models/fireworks";
|
|
import handleStreamData from "../sdk/handleStreamData";
|
|
import {GoogleChatSdk} from "../sdk/models/google";
|
|
import {XaiChatSdk} from "../sdk/models/xai";
|
|
import {CerebrasSdk} from "../sdk/models/cerebras";
|
|
import {CloudflareAISdk} from "../sdk/models/cloudflareAi";
|
|
|
|
export interface StreamParams {
|
|
env: Env;
|
|
openai: OpenAI;
|
|
messages: any[];
|
|
model: string;
|
|
systemPrompt: string;
|
|
preprocessedContext: any;
|
|
attachments: any[];
|
|
tools: any[];
|
|
disableWebhookGeneration: boolean;
|
|
maxTokens: number;
|
|
}
|
|
|
|
interface StreamHandlerParams {
|
|
controller: ReadableStreamDefaultController;
|
|
encoder: TextEncoder;
|
|
webhook?: { url: string, payload: unknown };
|
|
dynamicContext?: any;
|
|
}
|
|
|
|
const activeStreamType = types.model({
|
|
name: types.optional(types.string, ""),
|
|
maxTokens: types.optional(types.number, 0),
|
|
systemPrompt: types.optional(types.string, ""),
|
|
model: types.optional(types.string, ""),
|
|
messages: types.optional(types.array(types.frozen()), []),
|
|
attachments: types.optional(types.array(types.frozen()), []),
|
|
tools: types.optional(types.array(types.frozen()), []),
|
|
disableWebhookGeneration: types.optional(types.boolean, false)
|
|
});
|
|
|
|
const activeStreamsMap = types.map(
|
|
activeStreamType,
|
|
);
|
|
|
|
const ChatService = types
|
|
.model('ChatService', {
|
|
openAIApiKey: types.optional(types.string, ""),
|
|
openAIBaseURL: types.optional(types.string, ""),
|
|
activeStreams: types.optional(
|
|
activeStreamsMap,
|
|
{} // Correct initialization
|
|
),
|
|
maxTokens: types.number,
|
|
systemPrompt: types.string
|
|
})
|
|
.volatile(self => ({
|
|
openai: {} as OpenAI,
|
|
env: {} as Env,
|
|
webhookStreamActive: false
|
|
}))
|
|
.actions(self => {
|
|
// Helper functions
|
|
const createMessageInstance = (message: any) => {
|
|
if (typeof message.content === 'string') {
|
|
return Message.create({
|
|
role: message.role,
|
|
content: message.content,
|
|
});
|
|
}
|
|
if (Array.isArray(message.content)) {
|
|
const m = O1Message.create({
|
|
role: message.role,
|
|
content: message.content.map(item => ({
|
|
type: item.type,
|
|
text: item.text
|
|
})),
|
|
});
|
|
return m;
|
|
}
|
|
throw new Error('Unsupported message format');
|
|
};
|
|
|
|
|
|
const handleAgentProcess = async (
|
|
{controller, encoder, webhook, dynamicContext}: StreamHandlerParams
|
|
) => {
|
|
if (!webhook) return;
|
|
dynamicContext.append("\n## Agent Results\n~~~markdown\n");
|
|
|
|
for await (const chunk of self.streamAgentData({webhook})) {
|
|
controller.enqueue(encoder.encode(chunk));
|
|
dynamicContext.append(chunk);
|
|
}
|
|
|
|
dynamicContext.append("\n~~~\n");
|
|
ChatSdk.sendDoubleNewline(controller, encoder);
|
|
};
|
|
|
|
const createStreamParams = async (
|
|
streamConfig: any,
|
|
dynamicContext: any,
|
|
durableObject: any
|
|
): Promise<StreamParams> => {
|
|
return {
|
|
env: self.env,
|
|
openai: self.openai,
|
|
messages: streamConfig.messages.map(createMessageInstance),
|
|
model: streamConfig.model,
|
|
systemPrompt: streamConfig.systemPrompt,
|
|
preprocessedContext: getSnapshot(dynamicContext),
|
|
attachments: streamConfig.attachments ?? [],
|
|
tools: streamConfig.tools ?? [],
|
|
disableWebhookGeneration: true,
|
|
maxTokens: await durableObject.dynamicMaxTokens(
|
|
streamConfig.messages,
|
|
2000
|
|
),
|
|
}
|
|
};
|
|
|
|
const modelHandlers = {
|
|
openai: (params: StreamParams, dataHandler: Function) =>
|
|
OpenAiChatSdk.handleOpenAiStream(params, dataHandler),
|
|
groq: (params: StreamParams, dataHandler: Function) =>
|
|
GroqChatSdk.handleGroqStream(params, dataHandler),
|
|
claude: (params: StreamParams, dataHandler: Function) =>
|
|
ClaudeChatSdk.handleClaudeStream(params, dataHandler),
|
|
fireworks: (params: StreamParams, dataHandler: Function) =>
|
|
FireworksAiChatSdk.handleFireworksStream(params, dataHandler),
|
|
google: (params: StreamParams, dataHandler: Function) =>
|
|
GoogleChatSdk.handleGoogleStream(params, dataHandler),
|
|
xai: (params: StreamParams, dataHandler: Function) =>
|
|
XaiChatSdk.handleXaiStream(params, dataHandler),
|
|
cerebras: (params: StreamParams, dataHandler: Function) =>
|
|
CerebrasSdk.handleCerebrasStream(params, dataHandler),
|
|
cloudflareAI: (params: StreamParams, dataHandler: Function) =>
|
|
CloudflareAISdk.handleCloudflareAIStream(params, dataHandler)
|
|
};
|
|
|
|
return {
|
|
setActiveStream(streamId: string, stream: any) {
|
|
const validStream = {
|
|
name: stream?.name || "Unnamed Stream",
|
|
maxTokens: stream?.maxTokens || 0,
|
|
systemPrompt: stream?.systemPrompt || "",
|
|
model: stream?.model || "",
|
|
messages: stream?.messages || [],
|
|
attachments: stream?.attachments || [],
|
|
tools: stream?.tools || [],
|
|
disableWebhookGeneration: stream?.disableWebhookGeneration || false,
|
|
};
|
|
|
|
self.activeStreams.set(streamId, validStream);
|
|
},
|
|
|
|
removeActiveStream(streamId: string) {
|
|
self.activeStreams.delete(streamId);
|
|
},
|
|
setEnv(env: Env) {
|
|
self.env = env;
|
|
self.openai = new OpenAI({
|
|
apiKey: self.openAIApiKey,
|
|
baseURL: self.openAIBaseURL,
|
|
});
|
|
},
|
|
|
|
handleChatRequest: async (request: Request) => {
|
|
return ChatSdk.handleChatRequest(request, {
|
|
openai: self.openai,
|
|
env: self.env,
|
|
systemPrompt: self.systemPrompt,
|
|
maxTokens: self.maxTokens
|
|
});
|
|
},
|
|
|
|
setWebhookStreamActive(value) {
|
|
self.webhookStreamActive = value;
|
|
},
|
|
|
|
streamAgentData: async function* ({webhook}) {
|
|
console.log("streamAgentData::start");
|
|
if (self.webhookStreamActive) {
|
|
return
|
|
}
|
|
|
|
const queue: string[] = [];
|
|
let resolveQueueItem: Function;
|
|
let finished = false;
|
|
let errorOccurred: Error | null = null;
|
|
|
|
const dataPromise = () => new Promise<void>((resolve) => {
|
|
resolveQueueItem = resolve;
|
|
});
|
|
|
|
let currentPromise = dataPromise();
|
|
const eventSource = new EventSource(webhook.url.trim());
|
|
console.log("streamAgentData::setWebhookStreamActive::true");
|
|
self.setWebhookStreamActive(true)
|
|
try {
|
|
ChatSdk.handleAgentStream(eventSource, (data) => {
|
|
const formattedData = `data: ${JSON.stringify(data)}\n\n`;
|
|
queue.push(formattedData);
|
|
if (resolveQueueItem) resolveQueueItem();
|
|
currentPromise = dataPromise();
|
|
}).then(() => {
|
|
finished = true;
|
|
if (resolveQueueItem) resolveQueueItem();
|
|
}).catch((err) => {
|
|
console.log(`chatService::streamAgentData::STREAM_ERROR::${err}`);
|
|
errorOccurred = err;
|
|
if (resolveQueueItem) resolveQueueItem();
|
|
});
|
|
|
|
while (!finished || queue.length > 0) {
|
|
if (queue.length > 0) {
|
|
yield queue.shift()!;
|
|
} else if (errorOccurred) {
|
|
throw errorOccurred;
|
|
} else {
|
|
await currentPromise;
|
|
}
|
|
}
|
|
self.setWebhookStreamActive(false);
|
|
eventSource.close();
|
|
// console.log(`chatService::streamAgentData::complete`);
|
|
} catch (error) {
|
|
console.log(`chatService::streamAgentData::error`);
|
|
eventSource.close();
|
|
self.setWebhookStreamActive(false);
|
|
console.error("Error while streaming webhook data:", error);
|
|
throw error;
|
|
}
|
|
},
|
|
/**
|
|
* runModelHandler
|
|
* Selects the correct model handler and invokes it.
|
|
*/
|
|
async runModelHandler(params: {
|
|
streamConfig: any;
|
|
streamParams: any;
|
|
controller: ReadableStreamDefaultController;
|
|
encoder: TextEncoder;
|
|
streamId: string;
|
|
}) {
|
|
const {streamConfig, streamParams, controller, encoder, streamId} = params;
|
|
|
|
const modelFamily = getModelFamily(streamConfig.model);
|
|
console.log(
|
|
`chatService::handleSseStream::ReadableStream::modelFamily::${modelFamily}`
|
|
);
|
|
|
|
const handler = modelHandlers[modelFamily as ModelFamily];
|
|
if (handler) {
|
|
try {
|
|
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::start`);
|
|
await handler(streamParams, handleStreamData(controller, encoder));
|
|
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::finish`);
|
|
} catch (error) {
|
|
const message = error.message.toLowerCase();
|
|
|
|
if(message.includes("413 ") || (message.includes("maximum") || message.includes("too long") || message.includes("too large") )) {
|
|
throw new ClientError(`Error! Content length exceeds limits. Try shortening your message, removing any attached files, or editing an earlier message instead.`, 413, {model: streamConfig.model, maxTokens: streamParams.maxTokens})
|
|
// throw new Error(`Max tokens exceeded for model ${streamConfig.model}`)
|
|
}
|
|
if(message.includes("429 ")) {
|
|
throw new ClientError(`Error! Rate limit exceeded. Wait a few minutes before trying again.`, 429, {model: streamConfig.model, maxTokens: streamParams.maxTokens})
|
|
// throw new Error(`Max tokens exceeded for model ${streamConfig.model}`)
|
|
}
|
|
if (message.includes("404")) {
|
|
throw new ClientError(`Something went wrong, try again.`, 413, {})
|
|
// throw new Error(`Max tokens exceeded for model ${streamConfig.model}`)
|
|
}
|
|
throw error;
|
|
/*
|
|
'413 Request too large for model `mixtral-8x7b-32768` in organization `org_01htjxws48fm0rbbg5gnkgmbrh` service tier `on_demand` on tokens per minute (TPM): Limit 5000, Requested 49590, please reduce your message size and try again. Visit https://console.groq.com/docs/rate-limits for more information.'
|
|
*/
|
|
}
|
|
}
|
|
},
|
|
/**
|
|
* bootstrapAgents
|
|
* Checks if an agent exists, and if so, bootstraps it.
|
|
*/
|
|
async bootstrapAgents(params: {
|
|
savedStreamConfig: string;
|
|
controller: ReadableStreamDefaultController;
|
|
encoder: TextEncoder;
|
|
dynamicContext: any; // or more specific type
|
|
}) {
|
|
const {savedStreamConfig, controller, encoder, dynamicContext} = params;
|
|
|
|
const config = JSON.parse(savedStreamConfig);
|
|
const webhook = config?.webhooks?.[0];
|
|
|
|
if (webhook) {
|
|
console.log(`chatService::handleSseStream::ReadableStream::webhook:start`);
|
|
await handleAgentProcess({
|
|
controller,
|
|
encoder,
|
|
webhook,
|
|
dynamicContext,
|
|
});
|
|
console.log(`chatService::handleSseStream::ReadableStream::webhook::end`);
|
|
}
|
|
},
|
|
|
|
|
|
createSseReadableStream(params: {
|
|
streamId: string;
|
|
streamConfig: any;
|
|
savedStreamConfig: string;
|
|
durableObject: any;
|
|
}) {
|
|
const { streamId, streamConfig, savedStreamConfig, durableObject } = params;
|
|
|
|
return new ReadableStream({
|
|
async start(controller) {
|
|
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::open`);
|
|
const encoder = new TextEncoder();
|
|
|
|
try {
|
|
const dynamicContext = Message.create(streamConfig.preprocessedContext);
|
|
|
|
|
|
// Process agents if configured
|
|
await self.bootstrapAgents({
|
|
savedStreamConfig,
|
|
controller,
|
|
encoder,
|
|
dynamicContext: dynamicContext,
|
|
});
|
|
|
|
// Process the stream data using the appropriate handler
|
|
const streamParams = await createStreamParams(
|
|
streamConfig,
|
|
dynamicContext,
|
|
durableObject
|
|
);
|
|
|
|
try {
|
|
await self.runModelHandler({
|
|
streamConfig,
|
|
streamParams,
|
|
controller,
|
|
encoder,
|
|
streamId,
|
|
});
|
|
} catch (e) {
|
|
console.log("error caught at runModelHandler")
|
|
throw e;
|
|
}
|
|
|
|
} catch (error) {
|
|
console.error(`chatService::handleSseStream::${streamId}::Error`, error);
|
|
|
|
if(error instanceof ClientError) {
|
|
controller.enqueue(
|
|
encoder.encode(`data: ${JSON.stringify({ type: 'error', error: error.message })}\n\n`)
|
|
);
|
|
} else {
|
|
controller.enqueue(
|
|
encoder.encode(`data: ${JSON.stringify({ type: 'error', error: "Server error" })}\n\n`)
|
|
);
|
|
}
|
|
controller.close();
|
|
} finally {
|
|
try {
|
|
controller.close();
|
|
} catch (_) {}
|
|
}
|
|
},
|
|
});
|
|
},
|
|
|
|
|
|
handleSseStream: flow(function* (streamId: string): Generator<Promise<string>, Response, unknown> {
|
|
console.log(`chatService::handleSseStream::enter::${streamId}`);
|
|
|
|
// Check if a stream is already active for this ID
|
|
if (self.activeStreams.has(streamId)) {
|
|
console.log(`chatService::handleSseStream::${streamId}::[stream already active]`);
|
|
return new Response('Stream already active', { status: 409 });
|
|
}
|
|
|
|
// Retrieve the stream configuration from the durable object
|
|
const objectId = self.env.SITE_COORDINATOR.idFromName('stream-index');
|
|
const durableObject = self.env.SITE_COORDINATOR.get(objectId);
|
|
const savedStreamConfig = yield durableObject.getStreamData(streamId);
|
|
|
|
if (!savedStreamConfig) {
|
|
return new Response('Stream not found', { status: 404 });
|
|
}
|
|
|
|
const streamConfig = JSON.parse(savedStreamConfig);
|
|
console.log(`chatService::handleSseStream::${streamId}::[stream configured]`);
|
|
|
|
const stream = self.createSseReadableStream({
|
|
streamId,
|
|
streamConfig,
|
|
savedStreamConfig,
|
|
durableObject,
|
|
});
|
|
|
|
// Use `tee()` to create two streams: one for processing and one for the response
|
|
const [processingStream, responseStream] = stream.tee();
|
|
|
|
self.setActiveStream(streamId, {
|
|
...streamConfig,
|
|
});
|
|
|
|
processingStream.pipeTo(
|
|
new WritableStream({
|
|
close() {
|
|
self.removeActiveStream(streamId);
|
|
},
|
|
})
|
|
);
|
|
|
|
// Return the second stream as the response
|
|
return new Response(responseStream, {
|
|
headers: {
|
|
'Content-Type': 'text/event-stream',
|
|
'Cache-Control': 'no-cache',
|
|
'Connection': 'keep-alive',
|
|
},
|
|
});
|
|
}),
|
|
};
|
|
});
|
|
|
|
|
|
/**
|
|
* ClientError
|
|
* A custom construct for sending client-friendly errors via the controller in a structured and controlled manner.
|
|
*/
|
|
export class ClientError extends Error {
|
|
public statusCode: number;
|
|
public details: Record<string, any>;
|
|
|
|
constructor(message: string, statusCode: number, details: Record<string, any> = {}) {
|
|
super(message);
|
|
this.name = 'ClientError';
|
|
this.statusCode = statusCode;
|
|
this.details = details;
|
|
Object.setPrototypeOf(this, ClientError.prototype);
|
|
}
|
|
|
|
/**
|
|
* Formats the error for SSE-compatible data transmission.
|
|
*/
|
|
public formatForSSE(): string {
|
|
return JSON.stringify({
|
|
type: 'error',
|
|
message: this.message,
|
|
details: this.details,
|
|
statusCode: this.statusCode,
|
|
});
|
|
}
|
|
}
|
|
|
|
export default ChatService;
|