You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/content/docs/faq.md
+30Lines changed: 30 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -16,6 +16,36 @@ Here are answers to some of the most common questions.
16
16
17
17
Most gguf-based models should work, but newer models may require additions to the API. If a model doesn't work, please feel free to open up issues. However, be cautious about downloading models from the internet and directly onto your machine, as there may be security vulnerabilities in lama.cpp or ggml that could be maliciously exploited. Some models can be found on Hugging Face: https://huggingface.co/models?search=gguf, or models from gpt4all are compatible too: https://github.com/nomic-ai/gpt4all.
18
18
19
+
### Where are models stored?
20
+
21
+
LocalAI stores downloaded models in the following locations by default:
22
+
23
+
-**Command line**: `./models` (relative to current working directory)
24
+
-**Docker**: `/models` (inside the container, typically mounted to `./models` on host)
25
+
-**Launcher application**: `~/.localai/models` (in your home directory)
26
+
27
+
You can customize the model storage location using the `LOCALAI_MODELS_PATH` environment variable or `--models-path` command line flag. This is useful if you want to store models outside your home directory for backup purposes or to avoid filling up your home directory with large model files.
28
+
29
+
### How much storage space do models require?
30
+
31
+
Model sizes vary significantly depending on the model and quantization level:
- Ensure you have at least 2-3x the model size available for downloads and temporary files
44
+
- Use SSD storage for better performance
45
+
- Consider the model size relative to your system RAM - models larger than your RAM may not run efficiently
46
+
47
+
The WebUI shows model sizes in the Models tab to help you choose appropriate models for your system.
48
+
19
49
### Benchmarking LocalAI and llama.cpp shows different results!
20
50
21
51
LocalAI applies a set of defaults when loading models with the llama.cpp backend, one of these is mirostat sampling - while it achieves better results, it slows down the inference. You can disable this by setting `mirostat: 0` in the model config file. See also the advanced section ({{%relref "docs/advanced/advanced-usage" %}}) for more information and [this issue](https://github.com/mudler/LocalAI/issues/2780).
When starting LocalAI (either via Docker or via CLI) you can specify as argument a list of models to install automatically before starting the API, for example:
97
111
98
112
```bash
99
113
# From the model gallery (see available models with `local-ai models list`, in the WebUI from the model tab, or visiting https://models.localai.io)
@@ -112,7 +126,7 @@ local-ai run oci://localai/phi-2:latest
112
126
**Automatic Backend Detection**: When you install models from the gallery or YAML files, LocalAI automatically detects your system's GPU capabilities (NVIDIA, AMD, Intel) and downloads the appropriate backend. For advanced configuration options, see [GPU Acceleration]({{% relref "docs/features/gpu-acceleration#automatic-backend-detection" %}}).
113
127
{{% /alert %}}
114
128
115
-
For a full list of options, refer to the [Installer Options]({{% relref "docs/advanced/installer" %}}) documentation.
129
+
For a full list of options, you can run LocalAI with `--help` or refer to the [Installer Options]({{% relref "docs/advanced/installer" %}}) documentation.
116
130
117
131
Binaries can also be [manually downloaded]({{% relref "docs/reference/binaries" %}}).
0 commit comments