Skip to content

Text Generation Inference server #5934

Discussion options

You must be logged in to vote
  • We don't know yet if TGI will be supported in Haystack 2.0, but we are considering it (see #5625)

  • Falcon-40b-instruct seems to be supported by vLLM (its architecture should be FalconForCausalLM)

  • The article talks about Falcon-40b-instruct because at that time it was one of the best open-source LLMs.
    Today there are probably better open-source models (see Open LLM Leaderboard), which are also supported by vLLM.

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@Chance-Obondo
Comment options

@anakin87
Comment options

Answer selected by Chance-Obondo
@Chance-Obondo
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants