Files
open-web-agent-rs/README.md

28 lines
799 B
Markdown

# open-web-agent-rs
A Rust-based web agent with local inference capabilities.
## Components
### Local Inference Engine
The [Local Inference Engine](./local_inference_engine/README.md) provides a way to run large language models locally. It supports both CLI mode for direct text generation and server mode with an OpenAI-compatible API.
Features:
- Run Gemma models locally (1B, 2B, 7B, 9B variants)
- CLI mode for direct text generation
- Server mode with OpenAI-compatible API
- Support for various model configurations (base, instruction-tuned)
- Metal acceleration on macOS
See the [Local Inference Engine README](./local_inference_engine/README.md) for detailed usage instructions.
### Web Server
Server is being converted to MCP. Things are probably broken.
```text
bun i
bun dev
```