I noticed that llama-cpp has recently added support for chatglm3 and glm4, as detailed in [this pull request](https://github.com/ggerganov/llama.cpp/pull/8031). I hope that llama-cpp-python can also provide support for these models. Thank you for your continued hard work.