[Draft] GPT4all supported LLM architectures #53
Replies: 4 comments 12 replies
-
|
Supported quant types (from modellist.cpp@6518b33):
|
Beta Was this translation helpful? Give feedback.
-
|
Back to this topic again: XLlamaCPP is a Cython implementation that already provides builds for CUDA and Metal, with the latter enabled by default (see also setup.py#L74). What is missing is Vulkan for broad GPU support (xllamacpp#61). You could consider contributing your earlier Vulkan attempts to XLlamaCPP as they have already provided Metal builds. |
Beta Was this translation helpful? Give feedback.
-
|
This is an interesting project, I think the problem people are having with getting this up in running is that llama.cpp split their build up in a way which makes a single binary build per major platform, that works for everyone the way gpt4all was building (with baked in cuda + vulkan + kompute, etc), no longer tenable. so extensive use of --extra-index-url is needed for these backends. most projects I have seen do not cover all possible backends, though hosting the index on github is interesting |
Beta Was this translation helpful? Give feedback.
-
|
Looks like we might be getting XLlamaCPP for Vulkan (xllamacpp#61 (comment)), which would make it usable on non-NVIDIA Windows and Linux devices, supplementing the existing CUDA and MPS support. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
From llamamodel.cpp@6518b33:
Additionally, Mistral models work fine.
Beta Was this translation helpful? Give feedback.
All reactions