From 3c27cbfb9ab6a4f17fc25d7da75f669e10c0f153 Mon Sep 17 00:00:00 2001 From: Rafael dos Santos Silva Date: Tue, 26 Dec 2023 17:18:08 -0300 Subject: [PATCH] FIX: Use vLLM if TGI is not configured for OSS LLM inference (#380) --- lib/completions/endpoints/hugging_face.rb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/lib/completions/endpoints/hugging_face.rb b/lib/completions/endpoints/hugging_face.rb index f4290bae..22fd39f5 100644 --- a/lib/completions/endpoints/hugging_face.rb +++ b/lib/completions/endpoints/hugging_face.rb @@ -12,7 +12,7 @@ module DiscourseAi Llama2-chat-hf mistralai/Mixtral-8x7B-Instruct-v0.1 mistralai/Mistral-7B-Instruct-v0.2 - ].include?(model_name) + ].include?(model_name) && SiteSetting.ai_hugging_face_api_url.present? end def default_options