Where are the architectures implemented? #4328
-
Hello, I am very new to LLMs so please forgive my ignorance. I want to understand how these models are actually being implemented. So far, I understand that ggml helps with tensor ops, and the llama.cpp file has a lot of the workings of llama itself. Where are mistral and other llms implemented? Can we generate it from just the config file? Any help would be greatly appreciated! My goal is to understand how llama.cpp works! Cheers! |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 1 reply
-
it's all in llama.cpp, just a historic name |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
See https://github.com/ggerganov/llama.cpp/blob/e4b76bbe316ee50fb17d9ac29e654c0edf830eba/llama.cpp#L5584-L5627