fix(all): LLM templates on production.
This commit is contained in:
parent
6a5d6b37c2
commit
082b18a750
3 changed files with 16 additions and 3 deletions
|
@ -1391,7 +1391,9 @@ export class DialogKeywords {
|
|||
const conversation = min['apiConversations'][pid];
|
||||
const client = await GBUtil.getDirectLineClient(min);
|
||||
conversation.client = client;
|
||||
const response = await client.apis.Conversations.Conversations_StartConversation();
|
||||
const response = await client.apis.Conversations.Conversations_StartConversation(
|
||||
|
||||
);
|
||||
conversation.conversationId = response.obj.conversationId;
|
||||
|
||||
return await GBVMService.callVM('start', min, null, pid);
|
||||
|
|
|
@ -285,7 +285,7 @@ export class ChatServices {
|
|||
return { answer: undefined, questionId: 0 };
|
||||
}
|
||||
|
||||
const LLMMode = mode ?? answerMode;
|
||||
const LLMMode = (mode ?? answerMode).toLowerCase();
|
||||
const docsContext = min['vectorStore'];
|
||||
|
||||
let memory;
|
||||
|
@ -538,7 +538,6 @@ export class ChatServices {
|
|||
|
||||
GBLogEx.info(min, `Answer Source = ${answerSource}.`);
|
||||
|
||||
|
||||
let dataSource;
|
||||
if (answerSource === 'cache') {
|
||||
let sqliteFilePath =
|
||||
|
|
12
templates/llm-server.gbai/README.md
Normal file
12
templates/llm-server.gbai/README.md
Normal file
|
@ -0,0 +1,12 @@
|
|||
How to test (using api.pragmatismo.cloud as host):
|
||||
|
||||
POST https://api.pragmatismo.cloud/llmservergbot/dialogs/start
|
||||
|
||||
operator=123
|
||||
userSystemId=999
|
||||
|
||||
|
||||
POST https://api.pragmatismo.cloud/api/dk/messageBot
|
||||
|
||||
pid=1237189231897 (returned)
|
||||
text=soda
|
Loading…
Add table
Reference in a new issue