(improvement)(headless) Add support for the Ollama provider in the frontend and optimize the code (#1270)

This commit is contained in:
lexluo09
2024-06-28 17:29:58 +08:00
committed by GitHub
parent 7564256b0a
commit 528491717b
15 changed files with 165 additions and 62 deletions

View File

@@ -13,4 +13,25 @@ langchain4j:
embedding-model:
model-name: bge-small-zh
embedding-store:
persist-path: /tmp
persist-path: /tmp
# ollama:
# chat-model:
# base-url: http://localhost:11434
# api-key: demo
# model-name: qwen:0.5b
# temperature: 0.0
# timeout: PT60S
# chroma:
# embedding-store:
# baseUrl: http://0.0.0.0:8000
# timeout: 120s
# milvus:
# embedding-store:
# host: localhost
# port: 2379
# uri: http://0.0.0.0:19530
# token: demo
# dimension: 512
# timeout: 120s

View File

@@ -13,4 +13,25 @@ langchain4j:
embedding-model:
model-name: bge-small-zh
embedding-store:
persist-path: /tmp
persist-path: /tmp
# ollama:
# chat-model:
# base-url: http://localhost:11434
# api-key: demo
# model-name: qwen:0.5b
# temperature: 0.0
# timeout: PT60S
# chroma:
# embedding-store:
# baseUrl: http://0.0.0.0:8000
# timeout: 120s
# milvus:
# embedding-store:
# host: localhost
# port: 2379
# uri: http://0.0.0.0:19530
# token: demo
# dimension: 512
# timeout: 120s