File tree Expand file tree Collapse file tree 4 files changed +18
-16
lines changed Expand file tree Collapse file tree 4 files changed +18
-16
lines changed Original file line number Diff line number Diff line change 1- # Environment variables
2- LLM_API_KEY = dummy-key # This is fine for local node
1+ # Local Gaia node configuration
32LLM_API_BASE = http://localhost:8080/v1
4- LLM_MODEL = Qwen2.5-Coder-3B-Instruct
5- LLM_EMBED_MODEL = gte-Qwen2-1.5B-instruct
6- LLM_EMBED_SIZE = 1536
3+ LLM_MODEL = Qwen2.5-Coder-3B-Instruct-Q5_K_M
4+ LLM_EMBED_MODEL = local-embed
5+ LLM_EMBED_SIZE = 768
6+ # No API key needed for local setup
7+ LLM_API_KEY =
Original file line number Diff line number Diff line change @@ -12,11 +12,11 @@ jobs:
1212 runs-on : ubuntu-latest
1313
1414 env :
15- LLM_API_KEY : ${{ secrets.LLM_API_KEY }}
16- LLM_API_BASE : https://coder.gaia.domains/v1
17- LLM_MODEL : Qwen2.5-Coder-32B-Instruct-Q5_K_M
18- LLM_EMBED_MODEL : nomic-embed
15+ LLM_API_BASE : http://localhost:8080/v1
16+ LLM_MODEL : Qwen2.5-Coder-3B-Instruct-Q5_K_M
17+ LLM_EMBED_MODEL : local-embed
1918 LLM_EMBED_SIZE : 768
19+ LLM_API_KEY : " "
2020 SKIP_VECTOR_SEARCH : " true"
2121
2222 steps :
Original file line number Diff line number Diff line change 2121
2222app = FastAPI (title = "Rust Project Generator API" )
2323
24- # Get API key from environment variable
25- api_key = os .getenv ("LLM_API_KEY" )
26- if not api_key :
24+ # Get API key from environment variable (make optional)
25+ api_key = os .getenv ("LLM_API_KEY" , "" )
26+ # Only validate if not using local setup
27+ if not api_key and not os .getenv ("LLM_API_BASE" , "" ).startswith ("http://localhost" ):
2728 raise ValueError ("LLM_API_KEY environment variable not set" )
2829
2930# Get embedding size from environment variable
Original file line number Diff line number Diff line change @@ -5,10 +5,10 @@ services:
55 - " 8000:8000"
66 environment :
77 - LLM_API_KEY=${LLM_API_KEY}
8- - LLM_API_BASE=${LLM_API_BASE:-https://coder.gaia.domains /v1}
9- - LLM_MODEL=${LLM_MODEL:- Qwen2.5-Coder-32B -Instruct-Q5_K_M}
10- - LLM_EMBED_MODEL=${LLM_EMBED_MODEL:-nomic- embed}
11- - LLM_EMBED_SIZE=${LLM_EMBED_SIZE:- 768}
8+ - LLM_API_BASE=http://host.docker.internal:8080 /v1
9+ - LLM_MODEL=Qwen2.5-Coder-3B -Instruct-Q5_K_M
10+ - LLM_EMBED_MODEL=local- embed
11+ - LLM_EMBED_SIZE=768
1212 - QDRANT_HOST=qdrant
1313 - QDRANT_PORT=6333
1414 command : uvicorn app.main:app --host 0.0.0.0 --port 8000
You can’t perform that action at this time.
0 commit comments