Update README.md

Signed-off-by: Geoff Seemueller <28698553+geoffsee@users.noreply.github.com>
This commit is contained in:
Geoff Seemueller
2025-06-05 21:46:00 -04:00
committed by GitHub
parent bbc2120a84
commit 9a84675ffa

View File

@@ -1,27 +1,12 @@
# open-web-agent-rs # open-web-agent-rs
A Rust-based web agent with local inference capabilities. A Rust-based web agent with an embedded openai compatible inference server (supports gemma models only).
## Components ## Quickstart
```bash
### Local Inference Engine cp .env.example .env
The [Local Inference Engine](./local_inference_engine/README.md) provides a way to run large language models locally. It supports both CLI mode for direct text generation and server mode with an OpenAI-compatible API.
Features:
- Run Gemma models locally (1B, 2B, 7B, 9B variants)
- CLI mode for direct text generation
- Server mode with OpenAI-compatible API
- Support for various model configurations (base, instruction-tuned)
- Metal acceleration on macOS
See the [Local Inference Engine README](./local_inference_engine/README.md) for detailed usage instructions.
### Web Server
Server is being converted to MCP. Things are probably broken.
```text
bun i bun i
(cd local_inference_server && cargo run --release -- --server)
docker compose up -d searxng
bun dev bun dev
``` ```