From da71adea0970d6d950fb966d365cfb428aef8298 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Carlos=20Mochol=C3=AD?= Date: Wed, 19 Jul 2023 11:26:04 +0200 Subject: [PATCH] Link to Lit-GPT for LLaMA 2 (#427) --- README.md | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index fd31e36d..6ba807f3 100644 --- a/README.md +++ b/README.md @@ -24,7 +24,11 @@ This implementation builds on [nanoGPT](). The open-source code in this repository works with the original LLaMA weights that are distributed by Meta under a [research-only license](https://github.com/facebookresearch/llama/blob/main/MODEL_CARD.md#model-details). -New Apache 2.0 licensed weights are being released as part of the [Open LLaMA project](https://github.com/openlm-research/open_llama). To use the Open LLaMA weights or other LLaMA-like checkpoints such as Vicuna, check out the [Lit-GPT repository](https://github.com/Lightning-AI/lit-gpt). +## Looking for LLaMA 2? + +Meta AI has since released LLaMA 2. Additionally, new Apache 2.0 licensed weights are being released as part of the [Open LLaMA project](https://github.com/openlm-research/open_llama). + +To run LLaMA 2 weights, Open LLaMA weights, or Vicuna weights (among other LLaMA-like checkpoints), **check out the [Lit-GPT repository](https://github.com/Lightning-AI/lit-gpt)**. ## Why?