From 90f7a79380ff71eabeaea737d12a9eeea56df5ce Mon Sep 17 00:00:00 2001 From: lexluo09 <39718951+lexluo09@users.noreply.github.com> Date: Thu, 27 Jun 2024 14:27:46 +0800 Subject: [PATCH] (improvement)(config) Remove the supersonic-env configuration file and use the llm-config.yaml file uniformly. (#1238) --- assembly/bin/supersonic-daemon.bat | 1 - assembly/bin/supersonic-daemon.sh | 4 --- .../src/main/resources/application-local.yaml | 31 +------------------ .../src/main/resources/application.yaml | 5 ++- .../src/main/resources/llm-config.yaml | 16 ++++++++++ .../src/main/resources/supersonic-env.bat | 9 ------ .../src/main/resources/supersonic-env.sh | 10 ------ .../src/test/resources/application-local.yaml | 26 +--------------- .../src/test/resources/application.yaml | 5 ++- .../src/test/resources/llm-config.yaml | 16 ++++++++++ 10 files changed, 42 insertions(+), 81 deletions(-) create mode 100644 launchers/standalone/src/main/resources/llm-config.yaml delete mode 100644 launchers/standalone/src/main/resources/supersonic-env.bat delete mode 100644 launchers/standalone/src/main/resources/supersonic-env.sh create mode 100644 launchers/standalone/src/test/resources/llm-config.yaml diff --git a/assembly/bin/supersonic-daemon.bat b/assembly/bin/supersonic-daemon.bat index 2a003021a..f102569e5 100644 --- a/assembly/bin/supersonic-daemon.bat +++ b/assembly/bin/supersonic-daemon.bat @@ -4,7 +4,6 @@ chcp 65001 set "sbinDir=%~dp0" call %sbinDir%/supersonic-common.bat %* -call %sbinDir%/../conf/supersonic-env.bat %* set "command=%~1" set "service=%~2" diff --git a/assembly/bin/supersonic-daemon.sh b/assembly/bin/supersonic-daemon.sh index 618fe5229..641fae409 100755 --- a/assembly/bin/supersonic-daemon.sh +++ b/assembly/bin/supersonic-daemon.sh @@ -3,10 +3,6 @@ sbinDir=$(cd "$(dirname "$0")"; pwd) source $sbinDir/supersonic-common.sh -set -a -source $sbinDir/../conf/supersonic-env.sh -set +a - command=$1 service=$2 if [ -z "$service" ]; then diff --git a/launchers/standalone/src/main/resources/application-local.yaml b/launchers/standalone/src/main/resources/application-local.yaml index 449e2e06a..b22bead2a 100644 --- a/launchers/standalone/src/main/resources/application-local.yaml +++ b/launchers/standalone/src/main/resources/application-local.yaml @@ -85,33 +85,4 @@ swagger: name: email: url: '' - version: 3.0 - - -langchain4j: - open-ai: - chat-model: - # Replace with your LLM configs - # Note: The default API key `demo` is provided by langchain4j community - # which limits 1000 tokens per request. - base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} - api-key: ${OPENAI_API_KEY:demo} - model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo} - temperature: ${OPENAI_TEMPERATURE:0.0} - timeout: ${OPENAI_TIMEOUT:PT60S} -# embedding-model: -# base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} -# api-key: ${OPENAI_API_KEY:demo} -# dashscope: -# chat-model: -# api-key: ${OPENAI_API_KEY:demo} -# model-name: qwen-max-1201 -# embedding-model: -# api-key: ${OPENAI_API_KEY:demo} - in-memory: - embedding-model: - model-name: bge-small-zh - #modelPath: /data/model.onnx - #vocabularyPath: /data/onnx_vocab.txt - embedding-store: - file-path: /tmp + version: 3.0 \ No newline at end of file diff --git a/launchers/standalone/src/main/resources/application.yaml b/launchers/standalone/src/main/resources/application.yaml index 83b731d26..92c69b10e 100644 --- a/launchers/standalone/src/main/resources/application.yaml +++ b/launchers/standalone/src/main/resources/application.yaml @@ -3,5 +3,8 @@ spring: active: local application: name: chat + config: + import: + - classpath:llm-config.yaml mybatis: - mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml + mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml \ No newline at end of file diff --git a/launchers/standalone/src/main/resources/llm-config.yaml b/launchers/standalone/src/main/resources/llm-config.yaml new file mode 100644 index 000000000..39086c564 --- /dev/null +++ b/launchers/standalone/src/main/resources/llm-config.yaml @@ -0,0 +1,16 @@ +langchain4j: + open-ai: + chat-model: + # Replace with your LLM configs + # Note: The default API key `demo` is provided by langchain4j community + # which limits 1000 tokens per request. + base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} + api-key: ${OPENAI_API_KEY:demo} + model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo} + temperature: ${OPENAI_TEMPERATURE:0.0} + timeout: ${OPENAI_TIMEOUT:PT60S} + in-memory: + embedding-model: + model-name: bge-small-zh + embedding-store: + file-path: /tmp \ No newline at end of file diff --git a/launchers/standalone/src/main/resources/supersonic-env.bat b/launchers/standalone/src/main/resources/supersonic-env.bat deleted file mode 100644 index 8d6073358..000000000 --- a/launchers/standalone/src/main/resources/supersonic-env.bat +++ /dev/null @@ -1,9 +0,0 @@ - -rem Replace with your LLM configs -rem Note: The default API key `demo` is provided by langchain4j community -rem which limits 1000 tokens per request. -set OPENAI_API_BASE=http://langchain4j.dev/demo/openai/v1 -set OPENAI_API_KEY=demo -set OPENAI_MODEL_NAME=gpt-3.5-turbo -set OPENAI_TEMPERATURE=0.0 -set OPENAI_TIMEOUT=PT60S \ No newline at end of file diff --git a/launchers/standalone/src/main/resources/supersonic-env.sh b/launchers/standalone/src/main/resources/supersonic-env.sh deleted file mode 100644 index f6949a28a..000000000 --- a/launchers/standalone/src/main/resources/supersonic-env.sh +++ /dev/null @@ -1,10 +0,0 @@ -#!/usr/bin/env bash - -# Replace with your LLM configs -# Note: The default API key `demo` is provided by langchain4j community -# which limits 1000 tokens per request. -OPENAI_API_BASE=http://langchain4j.dev/demo/openai/v1 -OPENAI_API_KEY=demo -OPENAI_MODEL_NAME=gpt-3.5-turbo -OPENAI_TEMPERATURE=0.0 -OPENAI_TIMEOUT=PT60S diff --git a/launchers/standalone/src/test/resources/application-local.yaml b/launchers/standalone/src/test/resources/application-local.yaml index 659f188d5..18b3a3fcd 100644 --- a/launchers/standalone/src/test/resources/application-local.yaml +++ b/launchers/standalone/src/test/resources/application-local.yaml @@ -89,28 +89,4 @@ swagger: name: email: url: '' - version: 3.0 - - -langchain4j: - open-ai: - chat-model: - # Replace with your LLM configs - # Note: The default API key `demo` is provided by langchain4j community - # which limits 1000 tokens per request. - base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} - api-key: ${OPENAI_API_KEY:demo} - model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo} - temperature: ${OPENAI_TEMPERATURE:0.0} - timeout: ${OPENAI_TIMEOUT:PT60S} -# java.lang.RuntimeException: dev.ai4j.openai4j.OpenAiHttpException: Too many requests -# embedding-model: -# base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} -# api-key: ${OPENAI_API_KEY:demo} - in-memory: - embedding-model: - model-name: bge-small-zh - #modelPath: /data/model.onnx - #vocabularyPath: /data/onnx_vocab.txt - embedding-store: - file-path: /tmp + version: 3.0 \ No newline at end of file diff --git a/launchers/standalone/src/test/resources/application.yaml b/launchers/standalone/src/test/resources/application.yaml index 83b731d26..92c69b10e 100644 --- a/launchers/standalone/src/test/resources/application.yaml +++ b/launchers/standalone/src/test/resources/application.yaml @@ -3,5 +3,8 @@ spring: active: local application: name: chat + config: + import: + - classpath:llm-config.yaml mybatis: - mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml + mapper-locations=classpath:mappers/custom/*.xml,classpath*:/mappers/*.xml \ No newline at end of file diff --git a/launchers/standalone/src/test/resources/llm-config.yaml b/launchers/standalone/src/test/resources/llm-config.yaml new file mode 100644 index 000000000..39086c564 --- /dev/null +++ b/launchers/standalone/src/test/resources/llm-config.yaml @@ -0,0 +1,16 @@ +langchain4j: + open-ai: + chat-model: + # Replace with your LLM configs + # Note: The default API key `demo` is provided by langchain4j community + # which limits 1000 tokens per request. + base-url: ${OPENAI_API_BASE:https://api.openai.com/v1} + api-key: ${OPENAI_API_KEY:demo} + model-name: ${OPENAI_MODEL_NAME:gpt-3.5-turbo} + temperature: ${OPENAI_TEMPERATURE:0.0} + timeout: ${OPENAI_TIMEOUT:PT60S} + in-memory: + embedding-model: + model-name: bge-small-zh + embedding-store: + file-path: /tmp \ No newline at end of file