(improvement)(config) Remove the supersonic-env configuration file and use the llm-config.yaml file uniformly. (#1238)

This commit is contained in:
lexluo09
2024-06-27 14:27:46 +08:00
committed by GitHub
parent ff20ae4006
commit 90f7a79380
10 changed files with 42 additions and 81 deletions

View File

@@ -89,28 +89,4 @@ swagger:
name:
email:
url: ''
version: 3.0
langchain4j:
open-ai:
chat-model:
# Replace with your LLM configs
# Note: The default API key `demo` is provided by langchain4j community
# which limits 1000 tokens per request.
base-url: ${OPENAI_API_BASE:https://api.openai.com/v1}
api-key: ${OPENAI_API_KEY:demo}
model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo}
temperature: ${OPENAI_TEMPERATURE:0.0}
timeout: ${OPENAI_TIMEOUT:PT60S}
# java.lang.RuntimeException: dev.ai4j.openai4j.OpenAiHttpException: Too many requests
# embedding-model:
# base-url: ${OPENAI_API_BASE:https://api.openai.com/v1}
# api-key: ${OPENAI_API_KEY:demo}
in-memory:
embedding-model:
model-name: bge-small-zh
#modelPath: /data/model.onnx
#vocabularyPath: /data/onnx_vocab.txt
embedding-store:
file-path: /tmp
version: 3.0

View File

@@ -3,5 +3,8 @@ spring:
active: local
application:
name: chat
config:
import:
- classpath:llm-config.yaml
mybatis:
mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml
mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml

View File

@@ -0,0 +1,16 @@
langchain4j:
open-ai:
chat-model:
# Replace with your LLM configs
# Note: The default API key `demo` is provided by langchain4j community
# which limits 1000 tokens per request.
base-url: ${OPENAI_API_BASE:https://api.openai.com/v1}
api-key: ${OPENAI_API_KEY:demo}
model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo}
temperature: ${OPENAI_TEMPERATURE:0.0}
timeout: ${OPENAI_TIMEOUT:PT60S}
in-memory:
embedding-model:
model-name: bge-small-zh
embedding-store:
file-path: /tmp