Replies: 1 comment 2 replies
-
You should be able to change the value of |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I’m currently working on adapting LLaVA-style multimodal models to GGUF for efficient quantization and deployment. During this process, I encountered a persistent and deeply frustrating limitation related to the GGML_MAX_NAME constant.
Specifically, the 64-character tensor name limit seems to be hardcoded in a way that’s difficult to override externally. Despite updating GGML_MAX_NAME before including ggml.h, modifying relevant constants, and even rebuilding from source, the restriction persists—leading to truncated names or quantization failure due to duplicated, clipped identifiers.
This creates significant friction for research workflows, particularly in multimodal models like LLaVA and OpenFlamingo where tensor names naturally exceed 64 characters due to deeply nested submodules. It’s become a genuine blocker for clean, lossless GGUF conversion and downstream quantization.
Would you be willing to shed light on:
I completely understand the need for internal constraints in a performant library like ggml. That said, I and others in the community would greatly appreciate a clearer path or guidance for adapting the library to larger-scale, real-world models with complex naming hierarchies.
Beta Was this translation helpful? Give feedback.
All reactions