(improvement)(headless) Add support for the Ollama provider in the frontend and optimize the code (#1270)

This commit is contained in:
lexluo09
2024-06-28 17:29:58 +08:00
committed by GitHub
parent 7564256b0a
commit 528491717b
15 changed files with 165 additions and 62 deletions

View File

@@ -3,8 +3,8 @@ package com.tencent.supersonic.headless.chat.parser.llm;
import com.tencent.supersonic.common.config.LLMConfig;
import com.tencent.supersonic.headless.chat.query.llm.s2sql.LLMReq;
import com.tencent.supersonic.headless.chat.query.llm.s2sql.LLMResp;
import com.tencent.supersonic.common.util.S2ChatModelProvider;
import dev.langchain4j.model.chat.ChatLanguageModel;
import dev.langchain4j.model.provider.ChatLanguageModelProvider;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
import org.springframework.beans.factory.InitializingBean;
@@ -24,7 +24,7 @@ public abstract class SqlGenStrategy implements InitializingBean {
protected PromptHelper promptHelper;
protected ChatLanguageModel getChatLanguageModel(LLMConfig llmConfig) {
return S2ChatModelProvider.provide(llmConfig);
return ChatLanguageModelProvider.provide(llmConfig);
}
abstract LLMResp generate(LLMReq llmReq);