Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Linux] (Exit code 133) Error when loading large LLM models #285

Open
yushijinhun opened this issue Jan 6, 2025 · 40 comments
Open

[Linux] (Exit code 133) Error when loading large LLM models #285

yushijinhun opened this issue Jan 6, 2025 · 40 comments
Assignees
Labels
bug Something isn't working

Comments

@yushijinhun
Copy link

When loading large LLMs (for example, Meta-Llama-3.1-70B-Instruct-IQ2_S with context window 32768), I would encounter the error (Exit code: 133). Please check settings and try loading the model again.. My machine has 64G RAM and 16G vRAM, and I can load the model with same configuration with llama.cpp. Therefore, the problem should not be caused by insufficient RAM or vRAM.

Further investigation into the coredump shows that, the application crashes in function _ZN15partition_alloc8internal32PartitionExcessiveAllocationSizeEm, which means the application is trying to allocate an excessive amount of memory using PartitionAlloc in a call to posix_memalign. This explains why the problem occurs only on LM Studio but not on llama.cpp, as Electron uses PartitionAlloc by default. This is tracked in issue electron/electron#44291.

Apart from waiting for upstream to fix this issue, a potential workaround is to use a customized Electron build to disable PartitionAlloc.

@yagil
Copy link
Member

yagil commented Jan 6, 2025

@yushijinhun which operating system are you on? Assuming Linux?

@yushijinhun
Copy link
Author

@yushijinhun which operating system are you on? Assuming Linux?

Yes. I'm on Linux.

@yagil
Copy link
Member

yagil commented Jan 6, 2025

@yushijinhun which operating system are you on? Assuming Linux?

Yes. I'm on Linux.

Thanks. @neilmehta24 from our team is investigating this with priority as we speak.

@yagil yagil added the bug Something isn't working label Jan 6, 2025
@yagil yagil changed the title Error when loading large LLM models (Exit code 133) Error when loading large LLM models Jan 6, 2025
@yushijinhun
Copy link
Author

I built Electron v33.0.2 with use_allocator_shim = false and enable_backup_ref_ptr_support = false. After replacing LM Studio's electron with the one I built, the problem is solved.

@krones9000
Copy link

Could you give a little more specific detail about exactly how you resolved this? I'm in a similar situation, on linux, 16gbVRAM 128GB system RAM, could load models in Oobabooga but getting the 133 error in LM studio. Not sure where to start or how I would "build" Electron or what Electron is and how it relates to the Linux app image of LM Studio that I load.

@laushunyu
Copy link

I run LM Studio v0.3.6 on Ubuntu 24.04, with 7840HS CPU, 96GB RAM and 16GB GPU VRAM, encountered the same issue: small models can be loaded normally but large models cannot be loaded.

I would appreciate it if @yushijinhun share your compiled Electron version.

@yushijinhun
Copy link
Author

Not sure where to start or how I would "build" Electron or what Electron is and how it relates to the Linux app image of LM Studio that I load.

Electron is a UI framework built on Chromium used by LM Studio.

I would appreciate it if @yushijinhun share your compiled Electron version.

Here is my Electron v33.0.2 Linux x86_64 build (without allocator shim): Google Drive

You can also follow Electron's Build Instructions to build it yourself.

To replace LM Studio' stock Electron, first unpack the AppImage:

/path/to/LM-Studio-0.3.6-8-x64.AppImage --appimage-extract

Then you would see a squashfs-root directory. Go into it, extract the Electron distribution zip, and overwrite all existent files. Run the electron executable (you may need to add --no-sandbox flag), and you will see LM Studio starting.

@daxime
Copy link

daxime commented Jan 7, 2025

@yushijinhun Thank you for that :)

@CHesketh76
Copy link

CHesketh76 commented Jan 10, 2025

Why is this not set to default? LMstudio exclusively uses GGUF files which implies most users are limited on Vram.

@yagil
Copy link
Member

yagil commented Jan 10, 2025

This is a new bug in 0.3.6 because we updated our Electron version.
Our recommendation: consider staying on 0.3.5 until we fix it properly in a new release (it's in the works)

@CHesketh76
Copy link

Same issue when downgrading.

@aamir-gmail
Copy link

I am using linux version of LM studio 0 3.6 , with Ubuntu 22.04 LTS , I have two 3090 card and 256 GB RAM and AMD 24 Core CPU , for some reason the Qwen2 VL model not load 7B or 70B I get the following message (Exit code: 133). Please check settings and try loading the model again. I was able to load 7B version of the same model on the windows machine (0.3.6) with 64 GB RAM and 12 GB 4070 card , here neither 72B or 4B does not load my NVidia driver is 550, I thought it was a memory issues , them I tried LLAMA 70B , which loaded just fine , the GPU utilisation was around 15 GB each . Are there any logs collected by LM studion which I can share to help with this problem.

@yagil
Copy link
Member

yagil commented Jan 13, 2025

Thanks @aamir-gmail! We are aware of this issue and we are working on a fix. The recommendation is to stay on 0.3.5 until it's out. Get 0.3.5 from https://lmstudio.ai/download#beta-and-experimental-releases

@aamir-gmail
Copy link

aamir-gmail commented Jan 13, 2025 via email

@aamir-gmail
Copy link

aamir-gmail commented Jan 13, 2025 via email

@yagil yagil marked this as a duplicate of lmstudio-ai/lms#136 Jan 17, 2025
@yagil yagil marked this as a duplicate of #322 Jan 18, 2025
@alphacentrino
Copy link

I'm facing with the same issue. I have no GPU, but I have 32GB ram and intel core vPRO i7. I'm using latest cersion of LMS0.3.8-4

@codevski
Copy link

Same issue here even if I try offload the whole thing to RAM (mave more then enough) errors out. currently just dropping down a model size for the time being. Happy to share any logs that you need.

@pseudomo
Copy link

Same issue on Ubuntu 24.04 with 64Gb RAM. The model I am trying to load is 34.8 Gb bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF/DeepSeek-R1-Distill-Qwen-32B-Q8_0.gguf
The version of LM Studio is 0.3.8-4

@alphacentrino
Copy link

alphacentrino commented Jan 24, 2025

I resolved this issue by doing the following, In my case I need to first install 'vulkan' tools

  1. sudo apt install vulkan-tools
  2. Then start LMStudio and under settings--> Runtime(second tab on left) --> Configure RunTime(on Right): set this to 'CPU llama.cpp(limux)' from drop down.
    You should now see two runtimes under 'MyRuntImes'(Right Bottom), one with 'CPU llama.cpp(linux)' and another one ' Vulkan llama.cpp(linux)'

@Jachyme
Copy link

Jachyme commented Jan 24, 2025

Same problem. 64GB RAM, GeForce RTX 3060, Linux Mint 22.1 Cinnamon, Model 'DeepSeek R1 Distill Llama 70B Q4_K_L': Error loading model. (Exit code: 133)

But with Windows 11 on the same machine, with the same model copied from the Linux folder to the Windows folder, it loads and works without problems.

Both LM Studio 0.3.8 (Build 4)

alphacentrino: your solution doesn't work on my computer.

@yagil
Copy link
Member

yagil commented Jan 24, 2025

We are working on a fix, sorry for the wait. In the meantime, 0.3.5 should work: https://lmstudio.ai/download#beta-and-experimental-releases

@yagil yagil changed the title (Exit code 133) Error when loading large LLM models [Linux] (Exit code 133) Error when loading large LLM models Jan 24, 2025
@CHesketh76
Copy link

@yagil O.3.5 has the same issue for me. I have two desktops and same issue occurs on both machines with 0.3.5 and the recent release.

@alphacentrino
Copy link

Hi Jachyme,
mine is Ubuntu 20.04. After starting your LM studio, try loading the model, then as soon as you get error in LM studio, go back to command line and see the logs and you should be able to find what it's looking for. u can share your logs here, may be someone can help.

@dynamiccreator
Copy link

dynamiccreator commented Jan 25, 2025

I have the same issue with ubuntu 22.04 (128GB RAM) (4GB VRAM).

However it works in 0.3.5 (Build 2) if I use:

  • CPU runtime from the settings

  • set offload layers explicitly (bin symbol appears) to 0
    (GPU runtime might also work with this setting)

  • reduce batch evaluation size to 128 or lower

The trick with the reduce batch evaluation size shouldn't even be necessary as OP pointed out it all works with vanilla llama.cpp.

BUT:

However going to newer version (0.3.7 and 0.3.8) to use the R1 Distill models for example , the "trick" with the reduced batch eval size does not work anymore and it tells me:

🥲 Failed to load the model

Error loading model.

(Exit code: 133). Please check settings and try loading the model again. 

It loads the model and then at 10GB RAM usage or so the error occurs.

EDIT:

BUG LOGS:
With error (when context is higher, like even more than 5000 ) :

[ModelLoadingProvider] Requested to load model bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf with opts {
identifier: { desired: 'deepseek-r1-distill-qwen-14b', conflictBehavior: 'bump' },
instanceLoadTimeConfig: { fields: [ [Object], [Object], [Object], [Object] ] }
}
[ModelLoadingProvider] Started loading model bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf
[ModelProxyObject(id=deepseek-r1-distill-qwen-14b)] Forking LLMWorker with custom envVars: {}
22:59:31.212 › [LMSInternal][Client=LM Studio][Endpoint=loadModel] Error in channel handler: Error: Error loading model.
at _0x5bd2a1._0xf4108d (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:287003)
at _0x5bd2a1.emit (node:events:519:28)
at _0x5bd2a1.onChildExit (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:243719)
at ForkUtilityProcess. (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:243035)
at ForkUtilityProcess.emit (node:events:519:28)
at ForkUtilityProcess.a.emit (node:electron/js2c/browser_init:2:71438)
[LMSInternal][Client=LM Studio][Endpoint=loadModel] Error in loadModel channel _0x548ef3 [Error]: Error loading model.
at _0x5bd2a1._0xf4108d (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:287003)
at _0x5bd2a1.emit (node:events:519:28)
at _0x5bd2a1.onChildExit (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:243719)
at ForkUtilityProcess. (/tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:243035)
at ForkUtilityProcess.emit (node:events:519:28)
at ForkUtilityProcess.a.emit (node:electron/js2c/browser_init:2:71438) {
cause: '(Exit code: 133). Please check settings and try loading the model again. ',
suggestion: '',
errorData: undefined,
data: undefined,
displayData: undefined,
title: 'Error loading model.'
}
22:59:31.212 › [LMSInternal][Client=LM Studio][Endpoint=loadModel] No instance reference assigned before error
22:59:32.214 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735
22:59:32.214 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735
22:59:32.214 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735
22:59:32.214 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735
22:59:32.214 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735
22:59:32.214 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
at /tmp/.mount_LM-StuIorlnq/resources/app/.webpack/main/index.js:24:283735

With low context no problem:

[ModelLoadingProvider] Requested to load model bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf with opts {
identifier: { desired: 'deepseek-r1-distill-qwen-14b', conflictBehavior: 'bump' },
instanceLoadTimeConfig: { fields: [ [Object], [Object], [Object], [Object] ] }
}
[ModelLoadingProvider] Started loading model bartowski/DeepSeek-R1-Distill-Qwen-14B-GGUF/DeepSeek-R1-Distill-Qwen-14B-Q4_K_M.gguf
[ModelProxyObject(id=deepseek-r1-distill-qwen-14b)] Forking LLMWorker with custom envVars: {}

@aamir-gmail
Copy link

aamir-gmail commented Jan 26, 2025 via email

@dynamiccreator
Copy link

dynamiccreator commented Jan 26, 2025

I have edited my comment and added logs to my description. I hope these are the log messages you need.

@yagil yagil pinned this issue Jan 27, 2025
@yagil yagil unpinned this issue Jan 27, 2025
@yagil yagil pinned this issue Jan 27, 2025
@nukopal
Copy link

nukopal commented Jan 28, 2025

I have met same situation/errors, that says "(Exit code: 133). Please check settings and try loading the model again. ".
I will attach error logs, I hope help you to solve this problem.

error logs

in case of error occurs

[CachedFileDataProvider] Watching file at /home/nukopal-local/.lmstudio/conversations/1738044409708.conversation.json
[LMSInternal][Client=LM Studio][Endpoint=unloadModel] Unloading model KnNQkjVXd68ouJoIz5i1KQDU
Unloading model: KnNQkjVXd68ouJoIz5i1KQDU
15:46:42.668 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Cannot find model of instance reference. KnNQkjVXd68ouJoIz5i1KQDU
    at _0x6f0a0f.<computed>.getInstanceBySpecifierOrThrow (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:63:12348)
    at _0x6f0a0f.<computed>.getLLMModelBySpecifierOrThrow (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:63:14931)
    at Object.handler (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:68:19977)
    at _0x546cd8.<computed>.receivedRpcCall (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:365:147)
    at _0x41a188.receivedMessage (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:338:13495)
    at MessagePortMain.<anonymous> (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:338:11926)
    at MessagePortMain.emit (node:events:519:28)
    at MessagePortMain._internalPort.emit (node:electron/js2c/browser_init:2:105846)
[CachedFileDataProvider] Watching file at /home/nukopal-local/.lmstudio/.internal/user-concrete-model-default-config/.json
[CachedFileDataProvider] Watching file at /home/nukopal-local/.lmstudio/.internal/user-concrete-model-default-config/lmstudio-community/DeepSeek-R1-Distill-Llama-70B-GGUF/DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf.json
[ModelLoadingProvider] Requested to load model lmstudio-community/DeepSeek-R1-Distill-Llama-70B-GGUF/DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf with opts {
  identifier: {
    desired: 'deepseek-r1-distill-llama-70b',
    conflictBehavior: 'bump'
  },
  instanceLoadTimeConfig: { fields: [] }
}
[CachedFileDataProvider] Watching file at /home/nukopal-local/.lmstudio/.internal/user-concrete-model-default-config/lmstudio-community/DeepSeek-R1-Distill-Llama-70B-GGUF/DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf.json
[ModelLoadingProvider] Started loading model lmstudio-community/DeepSeek-R1-Distill-Llama-70B-GGUF/DeepSeek-R1-Distill-Llama-70B-Q4_K_M.gguf
[ModelProxyObject(id=deepseek-r1-distill-llama-70b)] Forking LLMWorker with custom envVars: {}
15:47:03.426 › [LMSInternal][Client=LM Studio][Endpoint=loadModel] Error in channel handler: Error: Error loading model.
    at _0x5bd2a1._0xf4108d (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:287003)
    at _0x5bd2a1.emit (node:events:519:28)
    at _0x5bd2a1.onChildExit (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:243719)
    at ForkUtilityProcess.<anonymous> (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:243035)
    at ForkUtilityProcess.emit (node:events:519:28)
    at ForkUtilityProcess.a.emit (node:electron/js2c/browser_init:2:71438)
[LMSInternal][Client=LM Studio][Endpoint=loadModel] Error in loadModel channel _0x548ef3 [Error]: Error loading model.
    at _0x5bd2a1._0xf4108d (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:287003)
    at _0x5bd2a1.emit (node:events:519:28)
    at _0x5bd2a1.onChildExit (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:243719)
    at ForkUtilityProcess.<anonymous> (/home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:243035)
    at ForkUtilityProcess.emit (node:events:519:28)
    at ForkUtilityProcess.a.emit (node:electron/js2c/browser_init:2:71438) {
  cause: '(Exit code: 133). Please check settings and try loading the model again. ',
  suggestion: '',
  errorData: undefined,
  data: undefined,
  displayData: undefined,
  title: 'Error loading model.'
}
15:47:03.430 › [LMSInternal][Client=LM Studio][Endpoint=loadModel] No instance reference assigned before error
15:47:04.426 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735
15:47:04.427 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735
15:47:04.428 › [LMSInternal][Client=LM Studio][Endpoint=countTokens] Error in RPC handler: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735
15:47:04.428 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735
15:47:04.429 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735
15:47:04.429 › Unhandled Rejection at: {} reason: Error: Model is unloaded.
    at /home/nukopal-local/.cache/LMStudio/.mount_LM-Stu3gPhtc/resources/app/.webpack/main/index.js:24:283735

in case of successful pattern

[ModelLoadingProvider] Requested to load model lmstudio-community/DeepSeek-R1-Distill-Llama-8B-GGUF/DeepSeek-R1-Distill-Llama-8B-Q8_0.gguf with opts {
  identifier: { desired: 'deepseek-r1-distill-llama-8b', conflictBehavior: 'bump' },
  instanceLoadTimeConfig: { fields: [] },
  instanceOperationTimeConfig: { fields: [] }
}
[CachedFileDataProvider] Watching file at /home/nukopal-local/.lmstudio/.internal/user-concrete-model-default-config/lmstudio-community/DeepSeek-R1-Distill-Llama-8B-GGUF/DeepSeek-R1-Distill-Llama-8B-Q8_0.gguf.json
[ModelLoadingProvider] Started loading model lmstudio-community/DeepSeek-R1-Distill-Llama-8B-GGUF/DeepSeek-R1-Distill-Llama-8B-Q8_0.gguf
[ModelProxyObject(id=deepseek-r1-distill-llama-8b)] Forking LLMWorker with custom envVars: {}

environment

Software

  • OS: Ubuntu 24.04.1 LTS (Deskjtop)
  • Kernel version : 6.8.0-51-generic #52-Ubuntu SMP PREEMPT_DYNAMIC Thu Dec 5 13:09:44 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux
  • LM Studio version: 0.3.8 (Build 4)
  • Runtime: CPU llama.cpp (Linux) v1.10.0
    (Vulkan llama.cpp v1.10.0 installed, but I don't use because machine only have poor GPU )

Hardware

  • CPU: Intel Xeon 5218 (single)
  • RAM: 256GB (Pmem Memory mode, normal DRAM 64GB used as Cache)
  • chassis: ThinkSystem P920

@dynamiccreator
Copy link

I want to add that it works with no problem if all layers + cache fits into VRAM. But with 0 layers to GPU it stops loading and that is not bound to the model size but only to the context size. So even a 70B model can be loaded if I set the context to 1000 or so.

@illtellyoulater
Copy link

illtellyoulater commented Jan 30, 2025

We are working on a fix, sorry for the wait. In the meantime, 0.3.5 should work: https://lmstudio.ai/download#beta-and-experimental-releases

@yagil any news on this?

By the way, latest 0.3.9 builds - both stable or beta - don't fix this and users are reporting neither the 0.3.5 you linked does.

Are there any cons to the proposed change in the Electron compilation flag discussed above, which seems to fix the issue?
Or would switching the flag interfere with other LM Studio features?

@Insurgent65
Copy link

Does not seem to be solved in the new version 0.3.9

@bijavix
Copy link

bijavix commented Jan 31, 2025

Version 0.3.9 still has the issue. By lowering the Context Length, the model was able to load.

@Msakhibullin24
Copy link

Version 0.3.9 still has the issue. By lowering the Context Length, the model was able to load.

how did you solve this problem?

I'm trying to run deepseekr1 gguf which at 20gpu would run

@Msakhibullin24
Copy link

how did you solve this problem?

I'm trying to run deepseekr1 gguf which at 20gpu would run

@krones9000
Copy link

Could you give a little more specific detail about exactly how you resolved this? I'm in a similar situation, on linux, 16gbVRAM 128GB system RAM, could load models in Oobabooga but getting the 133 error in LM studio. Not sure where to start or how I would "build" Electron or what Electron is and how it relates to the Linux app image of LM Studio that I load.

For what it's worth, I updated to the latest beta and that solved it for me.

@Insurgent65
Copy link

I have not been able to get it to work with any version or any of the tricks mentioned.

The user interface of the application is the best of all I've tried, it makes me very angry that I can't use it.

@mxfeinberg
Copy link

Electron's Build Instructions

I was running into this same issue over the last couple of days when attempting to use DeepSeek-R1 GGUF. This approach worked for me. I am now able to load the 300GB+ versions of the model.

@piotrkosecki
Copy link

0.3.9 build 6, issue still exists...

@fpatrick
Copy link

fpatrick commented Feb 4, 2025

I have same the problem on 0.3.9

@ellavs
Copy link

ellavs commented Feb 5, 2025

the same problem on 0.3.9, model qwen2.5-coder-7b-instruct. Ubuntu 20.04. Error:

Failed to load the model. Exit code 133. Please check settings and try loading the model again

@kP700c-github
Copy link

Update about models loaded successfully and unsuccessfully with LM Studio 0.3.9 build 6 (Linux Mint 22, AMD Ryzen 7 5700G, 64 GB RAM):

  • unsuccessful:
    • DeepSeek-R1-Distill-Qwen-32B-Q6_K.gguf
    • DeepSeek-R1-Distill-Qwen-32B-Q3_K_L.gguf
    • DeepSeek-Coder-V2-Lite-Instruct-Q8_0.gguf
    • FuseO1-DeekSeekR1-QwQ-SkyT1-32B-Preview-Q4_K_M.gguf
    • gemma-2-27b-it-Q4_K_M.gguf
    • gemma-2-9b-it-Q8_0.gguf
    • QwQ-32B-Preview-IQ3_M.gguf
    • qwen2.5-coder-32b-instruct-q4_0.gguf
    • qwen2.5-coder-14b-instruct-q4_0.gguf
    • Qwen2-VL-7B-Instruct-Q8_0.gguf+mmproj-Qwen2-VL-7B-Instruct-f16.gguf
    • Llama-3.3-70B-Instruct-Q3_K_L.gguf
  • successful:
    • DeepSeek-R1-Distill-Qwen-7B-Q8_0.gguf
    • DeepSeek-R1-Distill-Qwen-14B-Q8_0.gguf
    • Qwen2.5-7B-Instruct-Q8_0.gguf
    • Qwen2.5-7B-Instruct-1M-Q8_0.gguf
    • Qwen2.5-14B-Instruct-Q8_0.gguf
    • Qwen2.5-14B-Instruct-1M-Q8_0.gguf
    • Qwen2.5-Math-1.5B-Instruct-Q8_0.gguf
    • Dolphin3.0-Llama3.1-8B-Q8_0.gguf
    • phi-4-Q8_0.gguf
    • Yi-Coder-9B-Chat-Q8_0.gguf
    • Llama-3.2-4X3B-MOE-Ultra-Instruct-10B-D_AU-Q8_0.gguf
    • MiniCPM-V-2_6-Q8_0.gguf
    • MiniCPM-o-2_6-Q8_0.gguf
    • Mistral-Nemo-Instruct-2407-Q5_K_S.gguf
    • granite-3.1-8b-instruct-Q4_K_M.gguf
    • Codestral-22B-v0.1-Q4_K_M.gguf
    • gemma-2-2b-it-Q5_K_M.gguf
    • llama-3.2-3b-instruct-q8_0.gguf

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests