mirror of
https://github.com/geoffsee/open-gsio.git
synced 2025-09-08 22:56:46 +00:00
hotfix: add default for local-inference
This commit is contained in:
36
README.md
36
README.md
@@ -15,10 +15,10 @@
|
|||||||
- [Installation](#installation)
|
- [Installation](#installation)
|
||||||
- [Deployment](#deployment)
|
- [Deployment](#deployment)
|
||||||
- [Local Inference](#local-inference)
|
- [Local Inference](#local-inference)
|
||||||
|
- [mlx-omni-server (default)](#mlx-omni-server)
|
||||||
|
- [Adding models](#adding-models-for-local-inference-apple-silicon)
|
||||||
- [Ollama](#ollama)
|
- [Ollama](#ollama)
|
||||||
- [Adding models for local inference (ollama)](#adding-models-for-local-inference-ollama)
|
- [Adding models](#adding-models-for-local-inference-ollama)
|
||||||
- [mlx-omni-server (Apple Silicon Only)](#mlx-omni-server-apple-silicon-only)
|
|
||||||
- [Adding models for local inference (Apple Silicon)](#adding-models-for-local-inference-apple-silicon)
|
|
||||||
- [Testing](#testing)
|
- [Testing](#testing)
|
||||||
- [Troubleshooting](#troubleshooting)
|
- [Troubleshooting](#troubleshooting)
|
||||||
- [History](#history)
|
- [History](#history)
|
||||||
@@ -53,20 +53,9 @@
|
|||||||
|
|
||||||
## Local Inference
|
## Local Inference
|
||||||
> Local inference is achieved by overriding the `OPENAI_API_KEY` and `OPENAI_API_ENDPOINT` environment variables. See below.
|
> Local inference is achieved by overriding the `OPENAI_API_KEY` and `OPENAI_API_ENDPOINT` environment variables. See below.
|
||||||
### Ollama
|
|
||||||
~~~bash
|
|
||||||
bun run openai:local ollama # Start ollama server
|
|
||||||
bun run openai:local:enable # Configure connection
|
|
||||||
bun run server:dev # Restart server
|
|
||||||
~~~
|
|
||||||
#### Adding models for local inference (ollama)
|
|
||||||
|
|
||||||
~~~bash
|
### mlx-omni-server
|
||||||
# See https://ollama.com/library for available models
|
(default) (Apple Silicon Only) - Use Ollama for other platforms.
|
||||||
MODEL_TO_ADD=gemma3
|
|
||||||
docker exec -it ollama ollama run ${MODEL_TO_ADD}
|
|
||||||
~~~
|
|
||||||
### mlx-omni-server (Apple Silicon Only)
|
|
||||||
~~~bash
|
~~~bash
|
||||||
# (prereq) install mlx-omni-server
|
# (prereq) install mlx-omni-server
|
||||||
brew tap seemueller-io/tap
|
brew tap seemueller-io/tap
|
||||||
@@ -92,6 +81,21 @@ curl http://localhost:10240/v1/chat/completions \
|
|||||||
}"
|
}"
|
||||||
~~~
|
~~~
|
||||||
|
|
||||||
|
### Ollama
|
||||||
|
~~~bash
|
||||||
|
bun run openai:local ollama # Start ollama server
|
||||||
|
bun run openai:local:enable # Configure connection
|
||||||
|
bun run server:dev # Restart server
|
||||||
|
~~~
|
||||||
|
#### Adding models for local inference (ollama)
|
||||||
|
|
||||||
|
~~~bash
|
||||||
|
# See https://ollama.com/library for available models
|
||||||
|
MODEL_TO_ADD=gemma3
|
||||||
|
docker exec -it ollama ollama run ${MODEL_TO_ADD}
|
||||||
|
~~~
|
||||||
|
|
||||||
|
|
||||||
## Testing
|
## Testing
|
||||||
|
|
||||||
Tests are located in `__tests__` directories next to the code they test. Testing is incomplete at this time.
|
Tests are located in `__tests__` directories next to the code they test. Testing is incomplete at this time.
|
||||||
|
@@ -18,7 +18,7 @@
|
|||||||
"tail:email-service": "wrangler tail -c workers/email/wrangler-email.toml",
|
"tail:email-service": "wrangler tail -c workers/email/wrangler-email.toml",
|
||||||
"tail:analytics-service": "wrangler tail -c workers/analytics/wrangler-analytics.toml",
|
"tail:analytics-service": "wrangler tail -c workers/analytics/wrangler-analytics.toml",
|
||||||
"tail:session-proxy": "wrangler tail -c workers/session-proxy/wrangler-session-proxy.toml --env production",
|
"tail:session-proxy": "wrangler tail -c workers/session-proxy/wrangler-session-proxy.toml --env production",
|
||||||
"openai:local": "./scripts/start_inference_server.sh",
|
"openai:local": "./scripts/start_inference_server.sh mlx-omni-server",
|
||||||
"openai:local:mlx": "./scripts/start_inference_server.sh mlx-omni-server",
|
"openai:local:mlx": "./scripts/start_inference_server.sh mlx-omni-server",
|
||||||
"openai:local:ollama": "./scripts/start_inference_server.sh ollama",
|
"openai:local:ollama": "./scripts/start_inference_server.sh ollama",
|
||||||
"openai:local:configure": "scripts/configure_local_inference.sh",
|
"openai:local:configure": "scripts/configure_local_inference.sh",
|
||||||
|
Reference in New Issue
Block a user