mirror of
https://github.com/geoffsee/open-gsio.git
synced 2025-09-08 22:56:46 +00:00
fix build
This commit is contained in:
@@ -53,7 +53,7 @@ const createServerContext = (env, ctx) => {
|
||||
chatService: ChatService.create({
|
||||
openAIApiKey: env.OPENAI_API_KEY,
|
||||
openAIBaseURL: env.VITE_OPENAI_API_ENDPOINT,
|
||||
activeStreams: {},
|
||||
activeStreams: undefined,
|
||||
maxTokens: 16384,
|
||||
systemPrompt:
|
||||
"You are an assistant designed to provide accurate, concise, and context-aware responses while demonstrating your advanced reasoning capabilities.",
|
||||
|
@@ -27,6 +27,8 @@ export class AssistantSdk {
|
||||
tools
|
||||
.map((tool) => {
|
||||
switch (tool) {
|
||||
// case "user-attachments": return "### Attachments\nUser supplied attachments are normalized to text and will have this header (# Attachment:...) in the message.";
|
||||
// case "web-search": return "### Web Search\nResults are optionally available in 'Live Search'.";
|
||||
default:
|
||||
return `- ${tool}`;
|
||||
}
|
||||
|
@@ -1,22 +1,18 @@
|
||||
import { flow, getSnapshot, types } from "mobx-state-tree";
|
||||
import OpenAI from "openai";
|
||||
import ChatSdk from "../sdk/chat-sdk";
|
||||
import {getSnapshot, types} from 'mobx-state-tree';
|
||||
import OpenAI from 'openai';
|
||||
import ChatSdk from '../sdk/chat-sdk';
|
||||
import Message from "../models/Message";
|
||||
import O1Message from "../models/O1Message";
|
||||
import {
|
||||
getModelFamily,
|
||||
ModelFamily,
|
||||
} from "../../../src/components/chat/SupportedModels";
|
||||
import { OpenAiChatSdk } from "../sdk/models/openai";
|
||||
import { GroqChatSdk } from "../sdk/models/groq";
|
||||
import { ClaudeChatSdk } from "../sdk/models/claude";
|
||||
import { FireworksAiChatSdk } from "../sdk/models/fireworks";
|
||||
import {getModelFamily} from "../../../src/components/chat/SupportedModels";
|
||||
import {OpenAiChatSdk} from "../sdk/models/openai";
|
||||
import {GroqChatSdk} from "../sdk/models/groq";
|
||||
import {ClaudeChatSdk} from "../sdk/models/claude";
|
||||
import {FireworksAiChatSdk} from "../sdk/models/fireworks";
|
||||
import handleStreamData from "../sdk/handleStreamData";
|
||||
import { GoogleChatSdk } from "../sdk/models/google";
|
||||
import { XaiChatSdk } from "../sdk/models/xai";
|
||||
import { CerebrasSdk } from "../sdk/models/cerebras";
|
||||
import { CloudflareAISdk } from "../sdk/models/cloudflareAi";
|
||||
import {GoogleChatSdk} from "../sdk/models/google";
|
||||
import {XaiChatSdk} from "../sdk/models/xai";
|
||||
|
||||
// Types
|
||||
export interface StreamParams {
|
||||
env: Env;
|
||||
openai: OpenAI;
|
||||
@@ -33,486 +29,264 @@ export interface StreamParams {
|
||||
interface StreamHandlerParams {
|
||||
controller: ReadableStreamDefaultController;
|
||||
encoder: TextEncoder;
|
||||
webhook?: { url: string; payload: unknown };
|
||||
webhook?: { url: string, payload: unknown };
|
||||
dynamicContext?: any;
|
||||
}
|
||||
|
||||
const ChatService = types
|
||||
.model("ChatService", {
|
||||
openAIApiKey: types.optional(types.string, ""),
|
||||
openAIBaseURL: types.optional(types.string, ""),
|
||||
activeStreams: types.optional(
|
||||
types.map(
|
||||
types.model({
|
||||
name: types.optional(types.string, ""),
|
||||
maxTokens: types.optional(types.number, 0),
|
||||
systemPrompt: types.optional(types.string, ""),
|
||||
model: types.optional(types.string, ""),
|
||||
messages: types.optional(types.array(types.frozen()), []),
|
||||
attachments: types.optional(types.array(types.frozen()), []),
|
||||
tools: types.optional(types.array(types.frozen()), []),
|
||||
disableWebhookGeneration: types.optional(types.boolean, false),
|
||||
}),
|
||||
),
|
||||
),
|
||||
maxTokens: types.number,
|
||||
systemPrompt: types.string,
|
||||
})
|
||||
.volatile((self) => ({
|
||||
openai: {} as OpenAI,
|
||||
env: {} as Env,
|
||||
webhookStreamActive: false,
|
||||
}))
|
||||
.actions((self) => {
|
||||
const createMessageInstance = (message: any) => {
|
||||
if (typeof message.content === "string") {
|
||||
return Message.create({
|
||||
role: message.role,
|
||||
content: message.content,
|
||||
});
|
||||
}
|
||||
if (Array.isArray(message.content)) {
|
||||
const m = O1Message.create({
|
||||
role: message.role,
|
||||
content: message.content.map((item) => ({
|
||||
type: item.type,
|
||||
text: item.text,
|
||||
})),
|
||||
});
|
||||
return m;
|
||||
}
|
||||
throw new Error("Unsupported message format");
|
||||
};
|
||||
|
||||
const handleWebhookProcessing = async ({
|
||||
controller,
|
||||
encoder,
|
||||
webhook,
|
||||
dynamicContext,
|
||||
}: StreamHandlerParams) => {
|
||||
console.log("handleWebhookProcessing::start");
|
||||
if (!webhook) return;
|
||||
console.log("handleWebhookProcessing::[Loading Live Search]");
|
||||
dynamicContext.append("\n## Live Search\n~~~markdown\n");
|
||||
|
||||
for await (const chunk of self.streamWebhookData({ webhook })) {
|
||||
controller.enqueue(encoder.encode(chunk));
|
||||
dynamicContext.append(chunk);
|
||||
}
|
||||
|
||||
dynamicContext.append("\n~~~\n");
|
||||
console.log(
|
||||
`handleWebhookProcessing::[Finished loading Live Search!][length: ${dynamicContext.content.length}]`,
|
||||
);
|
||||
ChatSdk.sendDoubleNewline(controller, encoder);
|
||||
console.log("handleWebhookProcessing::exit");
|
||||
};
|
||||
|
||||
const createStreamParams = async (
|
||||
streamConfig: any,
|
||||
dynamicContext: any,
|
||||
durableObject: any,
|
||||
): Promise<StreamParams> => {
|
||||
return {
|
||||
env: self.env,
|
||||
openai: self.openai,
|
||||
messages: streamConfig.messages.map(createMessageInstance),
|
||||
model: streamConfig.model,
|
||||
systemPrompt: streamConfig.systemPrompt,
|
||||
preprocessedContext: getSnapshot(dynamicContext),
|
||||
attachments: streamConfig.attachments ?? [],
|
||||
tools: streamConfig.tools ?? [],
|
||||
disableWebhookGeneration: true,
|
||||
maxTokens: await durableObject.dynamicMaxTokens(
|
||||
streamConfig.messages,
|
||||
2000,
|
||||
),
|
||||
.model('ChatService', {
|
||||
openAIApiKey: types.optional(types.string, ""),
|
||||
openAIBaseURL: types.optional(types.string, ""),
|
||||
maxTokens: types.number,
|
||||
systemPrompt: types.string
|
||||
})
|
||||
.volatile(self => ({
|
||||
openai: {} as OpenAI,
|
||||
env: {} as Env,
|
||||
webhookStreamActive: false
|
||||
}))
|
||||
.actions(self => {
|
||||
// Helper functions
|
||||
const createMessageInstance = (message: any) => {
|
||||
if (typeof message.content === 'string') {
|
||||
return Message.create({
|
||||
role: message.role,
|
||||
content: message.content,
|
||||
});
|
||||
}
|
||||
if (Array.isArray(message.content)) {
|
||||
const m = O1Message.create({
|
||||
role: message.role,
|
||||
content: message.content.map(item => ({
|
||||
type: item.type,
|
||||
text: item.text
|
||||
})),
|
||||
});
|
||||
console.log({here: "createMessageInstance"});
|
||||
return m;
|
||||
}
|
||||
throw new Error('Unsupported message format');
|
||||
};
|
||||
};
|
||||
|
||||
const modelHandlers = {
|
||||
openai: (params: StreamParams, dataHandler: Function) =>
|
||||
OpenAiChatSdk.handleOpenAiStream(params, dataHandler),
|
||||
groq: (params: StreamParams, dataHandler: Function) =>
|
||||
GroqChatSdk.handleGroqStream(params, dataHandler),
|
||||
claude: (params: StreamParams, dataHandler: Function) =>
|
||||
ClaudeChatSdk.handleClaudeStream(params, dataHandler),
|
||||
fireworks: (params: StreamParams, dataHandler: Function) =>
|
||||
FireworksAiChatSdk.handleFireworksStream(params, dataHandler),
|
||||
google: (params: StreamParams, dataHandler: Function) =>
|
||||
GoogleChatSdk.handleGoogleStream(params, dataHandler),
|
||||
xai: (params: StreamParams, dataHandler: Function) =>
|
||||
XaiChatSdk.handleXaiStream(params, dataHandler),
|
||||
cerebras: (params: StreamParams, dataHandler: Function) =>
|
||||
CerebrasSdk.handleCerebrasStream(params, dataHandler),
|
||||
cloudflareAI: (params: StreamParams, dataHandler: Function) =>
|
||||
CloudflareAISdk.handleCloudflareAIStream(params, dataHandler),
|
||||
};
|
||||
|
||||
return {
|
||||
setActiveStream(streamId: string, stream: any) {
|
||||
const validStream = {
|
||||
name: stream?.name || "Unnamed Stream",
|
||||
maxTokens: stream?.maxTokens || 0,
|
||||
systemPrompt: stream?.systemPrompt || "",
|
||||
model: stream?.model || "",
|
||||
messages: stream?.messages || [],
|
||||
attachments: stream?.attachments || [],
|
||||
tools: stream?.tools || [],
|
||||
disableWebhookGeneration: stream?.disableWebhookGeneration || false,
|
||||
};
|
||||
const handleWebhookProcessing = async (
|
||||
{controller, encoder, webhook, dynamicContext}: StreamHandlerParams
|
||||
) => {
|
||||
console.log("handleWebhookProcessing::start");
|
||||
if (!webhook) return;
|
||||
console.log("handleWebhookProcessing::[Loading Hot Data]");
|
||||
dynamicContext.append("\n## Hot Data\n~~~markdown\n");
|
||||
|
||||
self.activeStreams.set(streamId, validStream);
|
||||
},
|
||||
|
||||
removeActiveStream(streamId: string) {
|
||||
self.activeStreams.delete(streamId);
|
||||
},
|
||||
setEnv(env: Env) {
|
||||
self.env = env;
|
||||
self.openai = new OpenAI({
|
||||
apiKey: self.openAIApiKey,
|
||||
baseURL: self.openAIBaseURL,
|
||||
});
|
||||
},
|
||||
|
||||
handleChatRequest: async (request: Request) => {
|
||||
return ChatSdk.handleChatRequest(request, {
|
||||
openai: self.openai,
|
||||
env: self.env,
|
||||
systemPrompt: self.systemPrompt,
|
||||
maxTokens: self.maxTokens,
|
||||
});
|
||||
},
|
||||
|
||||
setWebhookStreamActive(value) {
|
||||
self.webhookStreamActive = value;
|
||||
},
|
||||
|
||||
streamWebhookData: async function* ({ webhook }) {
|
||||
console.log("streamWebhookData::start");
|
||||
if (self.webhookStreamActive) {
|
||||
return;
|
||||
for await (const chunk of self.streamWebhookData({webhook})) {
|
||||
controller.enqueue(encoder.encode(chunk));
|
||||
dynamicContext.append(chunk);
|
||||
}
|
||||
|
||||
const queue: string[] = [];
|
||||
let resolveQueueItem: Function;
|
||||
let finished = false;
|
||||
let errorOccurred: Error | null = null;
|
||||
dynamicContext.append("\n~~~\n");
|
||||
console.log(`handleWebhookProcessing::[Finished loading Hot Data!][length: ${dynamicContext.content.length}]`);
|
||||
ChatSdk.sendDoubleNewline(controller, encoder);
|
||||
console.log("handleWebhookProcessing::exit")
|
||||
};
|
||||
|
||||
const dataPromise = () =>
|
||||
new Promise<void>((resolve) => {
|
||||
const createStreamParams = async (
|
||||
streamConfig: any,
|
||||
dynamicContext: any,
|
||||
durableObject: any
|
||||
): Promise<StreamParams> => {
|
||||
return {
|
||||
env: self.env,
|
||||
openai: self.openai,
|
||||
messages: streamConfig.messages.map(createMessageInstance),
|
||||
model: streamConfig.model,
|
||||
systemPrompt: streamConfig.systemPrompt,
|
||||
preprocessedContext: getSnapshot(dynamicContext),
|
||||
attachments: streamConfig.attachments ?? [],
|
||||
tools: streamConfig.tools ?? [],
|
||||
disableWebhookGeneration: true,
|
||||
maxTokens: await durableObject.dynamicMaxTokens(
|
||||
streamConfig.messages,
|
||||
128000
|
||||
),
|
||||
}
|
||||
};
|
||||
|
||||
const modelHandlers = {
|
||||
openai: (params: StreamParams, dataHandler: Function) =>
|
||||
OpenAiChatSdk.handleOpenAiStream(params, dataHandler),
|
||||
groq: (params: StreamParams, dataHandler: Function) =>
|
||||
GroqChatSdk.handleGroqStream(params, dataHandler),
|
||||
claude: (params: StreamParams, dataHandler: Function) =>
|
||||
ClaudeChatSdk.handleClaudeStream(params, dataHandler),
|
||||
fireworks: (params: StreamParams, dataHandler: Function) =>
|
||||
FireworksAiChatSdk.handleFireworksStream(params, dataHandler),
|
||||
google: (params: StreamParams, dataHandler: Function) =>
|
||||
GoogleChatSdk.handleGoogleStream(params, dataHandler),
|
||||
xai: (params: StreamParams, dataHandler: Function) =>
|
||||
XaiChatSdk.handleXaiStream(params, dataHandler),
|
||||
};
|
||||
|
||||
return {
|
||||
setEnv(env: Env) {
|
||||
self.env = env;
|
||||
self.openai = new OpenAI({
|
||||
apiKey: self.openAIApiKey,
|
||||
baseURL: self.openAIBaseURL,
|
||||
});
|
||||
},
|
||||
|
||||
handleChatRequest: async (request: Request) => {
|
||||
return ChatSdk.handleChatRequest(request, {
|
||||
openai: self.openai,
|
||||
env: self.env,
|
||||
systemPrompt: self.systemPrompt,
|
||||
maxTokens: self.maxTokens
|
||||
});
|
||||
},
|
||||
|
||||
setWebhookStreamActive(value) {
|
||||
self.webhookStreamActive = value;
|
||||
},
|
||||
|
||||
streamWebhookData: async function* ({webhook}) {
|
||||
console.log("streamWebhookData::start");
|
||||
if (self.webhookStreamActive) {
|
||||
return
|
||||
}
|
||||
|
||||
const queue: string[] = [];
|
||||
let resolveQueueItem: Function;
|
||||
let finished = false;
|
||||
let errorOccurred: Error | null = null;
|
||||
|
||||
const dataPromise = () => new Promise<void>((resolve) => {
|
||||
resolveQueueItem = resolve;
|
||||
});
|
||||
|
||||
let currentPromise = dataPromise();
|
||||
const eventSource = new EventSource(webhook.url.trim());
|
||||
console.log("streamWebhookData::setWebhookStreamActive::true");
|
||||
self.setWebhookStreamActive(true);
|
||||
try {
|
||||
ChatSdk.handleWebhookStream(eventSource, (data) => {
|
||||
const formattedData = `data: ${JSON.stringify(data)}\n\n`;
|
||||
queue.push(formattedData);
|
||||
if (resolveQueueItem) resolveQueueItem();
|
||||
currentPromise = dataPromise();
|
||||
})
|
||||
.then(() => {
|
||||
let currentPromise = dataPromise();
|
||||
const eventSource = new EventSource(webhook.url.trim());
|
||||
console.log("streamWebhookData::setWebhookStreamActive::true");
|
||||
self.setWebhookStreamActive(true)
|
||||
try {
|
||||
ChatSdk.handleWebhookStream(eventSource, (data) => {
|
||||
const formattedData = `data: ${JSON.stringify(data)}\n\n`;
|
||||
queue.push(formattedData);
|
||||
if (resolveQueueItem) resolveQueueItem();
|
||||
currentPromise = dataPromise();
|
||||
}).then(() => {
|
||||
finished = true;
|
||||
if (resolveQueueItem) resolveQueueItem();
|
||||
})
|
||||
.catch((err) => {
|
||||
console.log(
|
||||
`chatService::streamWebhookData::STREAM_ERROR::${err}`,
|
||||
);
|
||||
}).catch((err) => {
|
||||
console.log(`chatService::streamWebhookData::STREAM_ERROR::${err}`);
|
||||
errorOccurred = err;
|
||||
if (resolveQueueItem) resolveQueueItem();
|
||||
});
|
||||
|
||||
while (!finished || queue.length > 0) {
|
||||
if (queue.length > 0) {
|
||||
yield queue.shift()!;
|
||||
} else if (errorOccurred) {
|
||||
throw errorOccurred;
|
||||
} else {
|
||||
await currentPromise;
|
||||
}
|
||||
}
|
||||
self.setWebhookStreamActive(false);
|
||||
eventSource.close();
|
||||
console.log(`chatService::streamWebhookData::complete`);
|
||||
} catch (error) {
|
||||
console.log(`chatService::streamWebhookData::error`);
|
||||
eventSource.close();
|
||||
self.setWebhookStreamActive(false);
|
||||
console.error("Error while streaming webhook data:", error);
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
/**
|
||||
* runModelHandler
|
||||
* Selects the correct model handler and invokes it.
|
||||
*/
|
||||
async runModelHandler(params: {
|
||||
streamConfig: any;
|
||||
streamParams: any;
|
||||
controller: ReadableStreamDefaultController;
|
||||
encoder: TextEncoder;
|
||||
streamId: string;
|
||||
}) {
|
||||
const { streamConfig, streamParams, controller, encoder, streamId } =
|
||||
params;
|
||||
|
||||
const modelFamily = getModelFamily(streamConfig.model);
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::modelFamily::${modelFamily}`,
|
||||
);
|
||||
|
||||
const handler = modelHandlers[modelFamily as ModelFamily];
|
||||
if (handler) {
|
||||
try {
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::${streamId}::handler::start`,
|
||||
);
|
||||
await handler(streamParams, handleStreamData(controller, encoder));
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::${streamId}::handler::finish`,
|
||||
);
|
||||
} catch (error) {
|
||||
const message = error.message.toLowerCase();
|
||||
|
||||
if (
|
||||
message.includes("413 ") ||
|
||||
message.includes("maximum") ||
|
||||
message.includes("too long") ||
|
||||
message.includes("too large")
|
||||
) {
|
||||
throw new ClientError(
|
||||
`Error! Content length exceeds limits. Try shortening your message, removing any attached files, or editing an earlier message instead.`,
|
||||
413,
|
||||
{
|
||||
model: streamConfig.model,
|
||||
maxTokens: streamParams.maxTokens,
|
||||
},
|
||||
);
|
||||
}
|
||||
if (message.includes("429 ")) {
|
||||
throw new ClientError(
|
||||
`Error! Rate limit exceeded. Wait a few minutes before trying again.`,
|
||||
429,
|
||||
{
|
||||
model: streamConfig.model,
|
||||
maxTokens: streamParams.maxTokens,
|
||||
},
|
||||
);
|
||||
}
|
||||
if (message.includes("404")) {
|
||||
throw new ClientError(
|
||||
`Something went wrong, try again.`,
|
||||
413,
|
||||
{},
|
||||
);
|
||||
}
|
||||
throw error;
|
||||
/*
|
||||
'413 Request too large for model `mixtral-8x7b-32768` in organization `org_01htjxws48fm0rbbg5gnkgmbrh` service tier `on_demand` on tokens per minute (TPM): Limit 5000, Requested 49590, please reduce your message size and try again. Visit https://console.groq.com/docs/rate-limits for more information.'
|
||||
*/
|
||||
}
|
||||
}
|
||||
},
|
||||
/**
|
||||
* handleWebhookIfNeeded
|
||||
* Checks if a webhook exists, and if so, processes it.
|
||||
*/
|
||||
async handleWebhookIfNeeded(params: {
|
||||
savedStreamConfig: string;
|
||||
controller: ReadableStreamDefaultController;
|
||||
encoder: TextEncoder;
|
||||
}) {
|
||||
const { savedStreamConfig, controller, encoder, dynamicContext } =
|
||||
params;
|
||||
|
||||
const config = JSON.parse(savedStreamConfig);
|
||||
const webhook = config?.webhooks?.[0];
|
||||
|
||||
if (webhook) {
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::webhook:start`,
|
||||
);
|
||||
await handleWebhookProcessing({
|
||||
controller,
|
||||
encoder,
|
||||
webhook,
|
||||
dynamicContext,
|
||||
});
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::webhook::end`,
|
||||
);
|
||||
}
|
||||
},
|
||||
|
||||
createSseReadableStream(params: {
|
||||
streamId: string;
|
||||
streamConfig: any;
|
||||
savedStreamConfig: string;
|
||||
durableObject: any;
|
||||
}) {
|
||||
const { streamId, streamConfig, savedStreamConfig, durableObject } =
|
||||
params;
|
||||
|
||||
return new ReadableStream({
|
||||
async start(controller) {
|
||||
console.log(
|
||||
`chatService::handleSseStream::ReadableStream::${streamId}::open`,
|
||||
);
|
||||
const encoder = new TextEncoder();
|
||||
|
||||
try {
|
||||
const dynamicContext = Message.create(
|
||||
streamConfig.preprocessedContext,
|
||||
);
|
||||
|
||||
await self.handleWebhookIfNeeded({
|
||||
savedStreamConfig,
|
||||
controller,
|
||||
encoder,
|
||||
dynamicContext: dynamicContext,
|
||||
});
|
||||
|
||||
const streamParams = await createStreamParams(
|
||||
streamConfig,
|
||||
dynamicContext,
|
||||
durableObject,
|
||||
);
|
||||
|
||||
try {
|
||||
await self.runModelHandler({
|
||||
streamConfig,
|
||||
streamParams,
|
||||
controller,
|
||||
encoder,
|
||||
streamId,
|
||||
});
|
||||
} catch (e) {
|
||||
console.log("error caught at runModelHandler");
|
||||
throw e;
|
||||
}
|
||||
} catch (error) {
|
||||
console.error(
|
||||
`chatService::handleSseStream::${streamId}::Error`,
|
||||
error,
|
||||
);
|
||||
|
||||
if (error instanceof ClientError) {
|
||||
controller.enqueue(
|
||||
encoder.encode(
|
||||
`data: ${JSON.stringify({ type: "error", error: error.message })}\n\n`,
|
||||
),
|
||||
);
|
||||
while (!finished || queue.length > 0) {
|
||||
if (queue.length > 0) {
|
||||
yield queue.shift()!;
|
||||
} else if (errorOccurred) {
|
||||
throw errorOccurred;
|
||||
} else {
|
||||
controller.enqueue(
|
||||
encoder.encode(
|
||||
`data: ${JSON.stringify({ type: "error", error: "Server error" })}\n\n`,
|
||||
),
|
||||
);
|
||||
await currentPromise;
|
||||
}
|
||||
controller.close();
|
||||
} finally {
|
||||
try {
|
||||
controller.close();
|
||||
} catch (_) {}
|
||||
}
|
||||
},
|
||||
});
|
||||
},
|
||||
self.setWebhookStreamActive(false);
|
||||
eventSource.close();
|
||||
console.log(`chatService::streamWebhookData::complete`);
|
||||
} catch (error) {
|
||||
console.log(`chatService::streamWebhookData::error`);
|
||||
eventSource.close();
|
||||
self.setWebhookStreamActive(false);
|
||||
console.error("Error while streaming webhook data:", error);
|
||||
throw error;
|
||||
}
|
||||
},
|
||||
|
||||
handleSseStream: flow(function* (
|
||||
streamId: string,
|
||||
): Generator<Promise<string>, Response, unknown> {
|
||||
console.log(`chatService::handleSseStream::enter::${streamId}`);
|
||||
async handleSseStream(streamId: string) {
|
||||
console.log(`chatService::handleSseStream::enter::${streamId}`);
|
||||
|
||||
if (self.activeStreams.has(streamId)) {
|
||||
console.log(
|
||||
`chatService::handleSseStream::${streamId}::[stream already active]`,
|
||||
);
|
||||
return new Response("Stream already active", { status: 409 });
|
||||
}
|
||||
const objectId = self.env.SITE_COORDINATOR.idFromName('stream-index');
|
||||
const durableObject = self.env.SITE_COORDINATOR.get(objectId);
|
||||
const savedStreamConfig = await durableObject.getStreamData(streamId);
|
||||
// console.log({savedStreamConfig});
|
||||
|
||||
const objectId = self.env.SITE_COORDINATOR.idFromName("stream-index");
|
||||
const durableObject = self.env.SITE_COORDINATOR.get(objectId);
|
||||
const savedStreamConfig = yield durableObject.getStreamData(streamId);
|
||||
if (!savedStreamConfig) {
|
||||
return new Response('Stream not found', {status: 404});
|
||||
}
|
||||
|
||||
if (!savedStreamConfig) {
|
||||
return new Response("Stream not found", { status: 404 });
|
||||
}
|
||||
const streamConfig = JSON.parse(savedStreamConfig);
|
||||
console.log(`chatService::handleSseStream::${streamId}::[stream configured]`);
|
||||
|
||||
const streamConfig = JSON.parse(savedStreamConfig);
|
||||
console.log(
|
||||
`chatService::handleSseStream::${streamId}::[stream configured]`,
|
||||
);
|
||||
console.log(`chatService::handleSseStream::${streamId}::[initializing stream]`);
|
||||
|
||||
const stream = self.createSseReadableStream({
|
||||
streamId,
|
||||
streamConfig,
|
||||
savedStreamConfig,
|
||||
durableObject,
|
||||
});
|
||||
const stream = new ReadableStream({
|
||||
async start(controller) {
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::open`);
|
||||
const encoder = new TextEncoder();
|
||||
// controller.enqueue(encoder.encode('retry: 3\n\n'));
|
||||
|
||||
const [processingStream, responseStream] = stream.tee();
|
||||
console.log(streamConfig.preprocessedContext);
|
||||
|
||||
self.setActiveStream(streamId, {});
|
||||
const dynamicContext = Message.create(streamConfig.preprocessedContext);
|
||||
console.log(`chatService::handleSseStream::ReadableStream::dynamicContext`);
|
||||
try {
|
||||
|
||||
processingStream.pipeTo(
|
||||
new WritableStream({
|
||||
close() {
|
||||
console.log(
|
||||
`chatService::handleSseStream::${streamId}::[stream closed]`,
|
||||
);
|
||||
const config = JSON.parse(savedStreamConfig);
|
||||
|
||||
const webhook = config?.webhooks?.at(0);
|
||||
|
||||
if (webhook) {
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::webhook:start`);
|
||||
await handleWebhookProcessing({
|
||||
controller,
|
||||
encoder,
|
||||
webhook,
|
||||
dynamicContext
|
||||
});
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::webhook::end`);
|
||||
}
|
||||
|
||||
const streamParams = await createStreamParams(
|
||||
streamConfig,
|
||||
dynamicContext,
|
||||
durableObject
|
||||
);
|
||||
|
||||
const modelFamily = getModelFamily(streamConfig.model);
|
||||
console.log(`chatService::handleSseStream::ReadableStream::modelFamily::${modelFamily}`);
|
||||
const handler = modelHandlers[modelFamily];
|
||||
|
||||
if (handler) {
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::start`);
|
||||
await handler(
|
||||
streamParams,
|
||||
handleStreamData(controller, encoder)
|
||||
);
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::handler::finish`);
|
||||
}
|
||||
} catch (error) {
|
||||
controller.enqueue(
|
||||
encoder.encode(`data: ${JSON.stringify({error: error.message})}\n\n`)
|
||||
);
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::Error::${error}`);
|
||||
} finally {
|
||||
console.log(`chatService::handleSseStream::ReadableStream::${streamId}::closed::${streamId}`);
|
||||
controller.close();
|
||||
}
|
||||
},
|
||||
}),
|
||||
);
|
||||
|
||||
return new Response(responseStream, {
|
||||
headers: {
|
||||
"Content-Type": "text/event-stream",
|
||||
"Cache-Control": "no-cache",
|
||||
Connection: "keep-alive",
|
||||
},
|
||||
});
|
||||
}),
|
||||
};
|
||||
});
|
||||
|
||||
/**
|
||||
* ClientError
|
||||
* A custom construct for sending client-friendly errors via the controller in a structured and controlled manner.
|
||||
*/
|
||||
export class ClientError extends Error {
|
||||
public statusCode: number;
|
||||
public details: Record<string, any>;
|
||||
|
||||
constructor(
|
||||
message: string,
|
||||
statusCode: number,
|
||||
details: Record<string, any> = {},
|
||||
) {
|
||||
super(message);
|
||||
this.name = "ClientError";
|
||||
this.statusCode = statusCode;
|
||||
this.details = details;
|
||||
Object.setPrototypeOf(this, ClientError.prototype);
|
||||
}
|
||||
|
||||
/**
|
||||
* Formats the error for SSE-compatible data transmission.
|
||||
*/
|
||||
public formatForSSE(): string {
|
||||
return JSON.stringify({
|
||||
type: "error",
|
||||
message: this.message,
|
||||
details: this.details,
|
||||
statusCode: this.statusCode,
|
||||
});
|
||||
return new Response(
|
||||
stream,
|
||||
{
|
||||
headers: {
|
||||
'Content-Type': 'text/event-stream',
|
||||
'Cache-Control': 'no-cache',
|
||||
'Connection': 'keep-alive',
|
||||
},
|
||||
}
|
||||
);
|
||||
}
|
||||
};
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
export default ChatService;
|
||||
export default ChatService;
|
Reference in New Issue
Block a user