add openai compatible endpoint for chat completions
This commit is contained in:

committed by
Geoff Seemueller

parent
3b4c8b045a
commit
8a3c0797c3
21
README.md
21
README.md
@@ -1,10 +1,27 @@
|
||||
# open-web-agent-rs
|
||||
|
||||
A Rust-based web agent with local inference capabilities.
|
||||
|
||||
## Components
|
||||
|
||||
### Local Inference Engine
|
||||
|
||||
The [Local Inference Engine](./local_inference_engine/README.md) provides a way to run large language models locally. It supports both CLI mode for direct text generation and server mode with an OpenAI-compatible API.
|
||||
|
||||
Features:
|
||||
- Run Gemma models locally (1B, 2B, 7B, 9B variants)
|
||||
- CLI mode for direct text generation
|
||||
- Server mode with OpenAI-compatible API
|
||||
- Support for various model configurations (base, instruction-tuned)
|
||||
- Metal acceleration on macOS
|
||||
|
||||
See the [Local Inference Engine README](./local_inference_engine/README.md) for detailed usage instructions.
|
||||
|
||||
### Web Server
|
||||
|
||||
Server is being converted to MCP. Things are probably broken.
|
||||
|
||||
```text
|
||||
bun i
|
||||
bun dev
|
||||
```
|
||||
|
||||
|
||||
|
Reference in New Issue
Block a user