diff --git a/LLMFarm/Chats/ChatView.swift b/LLMFarm/Chats/ChatView.swift index 2ef6306..f69a628 100644 --- a/LLMFarm/Chats/ChatView.swift +++ b/LLMFarm/Chats/ChatView.swift @@ -132,7 +132,7 @@ struct ChatView: View { HStack{ Text(String(describing: aiChatModel.state)) .foregroundColor(.white) - .frame(width: 100 /*,height: 25*/) + .frame(width: 200 /*,height: 25*/) // .padding([.top, .leading], 5) .opacity(0.4) .offset(x: 0,y: 8) diff --git a/README.md b/README.md index a3c8010..9b7de8e 100644 --- a/README.md +++ b/README.md @@ -73,7 +73,7 @@ See full list [here](https://github.com/ggerganov/llama.cpp). - [x] [Yi-VL](https://huggingface.co/models?search=Yi-VL) - [x] [Moondream](https://huggingface.co/vikhyatk/moondream2) -Note: For *Falcon, Alpaca, GPT4All, Chinese LLaMA / Alpaca and Chinese LLaMA-2 / Alpaca-2, Vigogne (French), Vicuna, Koala, OpenBuddy (Multilingual), Pygmalion/Metharme, WizardLM, Baichuan 1 & 2 + derivations, Aquila 1 & 2, Mistral AI v0.1, Refact, Persimmon 8B, MPT, Bloom* select `llama inferece` in model settings. +Note: For *Falcon, Alpaca, GPT4All, Chinese LLaMA / Alpaca and Chinese LLaMA-2 / Alpaca-2, Vigogne (French), Vicuna, Koala, OpenBuddy (Multilingual), Pygmalion/Metharme, WizardLM, Baichuan 1 & 2 + derivations, Aquila 1 & 2, Mistral AI v0.1, Refact, Persimmon 8B, MPT, Bloom* select `llama inference` in model settings. # Sampling methods - [x] Temperature (temp, tok-k, top-p) diff --git a/llmfarm_core.swift b/llmfarm_core.swift index a6aab90..a5d543c 160000 --- a/llmfarm_core.swift +++ b/llmfarm_core.swift @@ -1 +1 @@ -Subproject commit a6aab90f35e11632b7bada26d74dff913ab7586a +Subproject commit a5d543cbac92003d4232dc4c26e372ca9fc67abc