diff --git a/packages/backend/src/managers/podmanConnection.ts b/packages/backend/src/managers/podmanConnection.ts index fec927b87..552c3d782 100644 --- a/packages/backend/src/managers/podmanConnection.ts +++ b/packages/backend/src/managers/podmanConnection.ts @@ -156,7 +156,7 @@ export class PodmanConnection extends Publisher s === vmtype); if (type === undefined) { return VMType.UNKNOWN; } diff --git a/packages/backend/src/workers/provider/LlamaCppPython.ts b/packages/backend/src/workers/provider/LlamaCppPython.ts index 183a6c621..56bda2b68 100644 --- a/packages/backend/src/workers/provider/LlamaCppPython.ts +++ b/packages/backend/src/workers/provider/LlamaCppPython.ts @@ -250,12 +250,10 @@ export class LlamaCppPython extends InferenceProvider { case VMType.WSL: return gpu?.vendor === GPUVendor.NVIDIA ? llamacpp.cuda : llamacpp.default; case VMType.LIBKRUN: + case VMType.LIBKRUN_LABEL: return gpu ? llamacpp.vulkan : llamacpp.default; // no GPU support - case VMType.QEMU: - case VMType.APPLEHV: - case VMType.HYPERV: - case VMType.UNKNOWN: + default: return llamacpp.default; } } diff --git a/packages/shared/src/models/IPodman.ts b/packages/shared/src/models/IPodman.ts index 9879f1ba3..2212eb749 100644 --- a/packages/shared/src/models/IPodman.ts +++ b/packages/shared/src/models/IPodman.ts @@ -19,8 +19,10 @@ export enum VMType { WSL = 'wsl', LIBKRUN = 'libkrun', + LIBKRUN_LABEL = 'GPU enabled (LibKrun)', QEMU = 'qemu', APPLEHV = 'applehv', + APPLEHV_LABEL = 'default (Apple HyperVisor)', HYPERV = 'hyperv', UNKNOWN = 'unknown', }