(improvement)(config) Split the application-local.yml file based on functionality. (#1250)

This commit is contained in:
lexluo09
2024-06-27 16:56:16 +08:00
committed by GitHub
parent 8b88528938
commit bbd61ac937
23 changed files with 162 additions and 163 deletions

View File

@@ -0,0 +1,16 @@
langchain4j:
open-ai:
chat-model:
# Replace with your LLM configs
# Note: The default API key `demo` is provided by langchain4j community
# which limits 1000 tokens per request.
base-url: ${OPENAI_API_BASE:https://api.openai.com/v1}
api-key: ${OPENAI_API_KEY:demo}
model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo}
temperature: ${OPENAI_TEMPERATURE:0.0}
timeout: ${OPENAI_TIMEOUT:PT60S}
in-memory:
embedding-model:
model-name: bge-small-zh
embedding-store:
file-path: /tmp