Text Generation Inference server #5934
-
Is there any or there any plans to do a custom prompt node integration to https://github.com/huggingface/text-generation-inference by huggingface which I have used to host the ✨ Falcon-40B-Instruct model. I have checked out Langchain and they have such a solution but there is not one by deepset haystack and this might force me to switch to Langchain which I really dont want to |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
Hello, @Chance-Obondo! We are definitely looking at TGI with interest. Other options:
I hope one of these options can help you... |
Beta Was this translation helpful? Give feedback.
We don't know yet if TGI will be supported in Haystack 2.0, but we are considering it (see #5625)
Falcon-40b-instruct
seems to be supported by vLLM (its architecture should beFalconForCausalLM
)The article talks about
Falcon-40b-instruct
because at that time it was one of the best open-source LLMs.Today there are probably better open-source models (see Open LLM Leaderboard), which are also supported by vLLM.