Replies: 2 comments
-
MoE - Mixture of experts (MoE) - https://en.wikipedia.org/wiki/Mixture_of_experts For those of us who are not just C/C++ or ML experts... |
Beta Was this translation helpful? Give feedback.
0 replies
-
We have implemented a similar approach called MixLoRA: TUDB-Labs/mLoRA#136 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
stupid question ahead
i want to do experiment with MoE but using lora adapter kind thing. My understanding is currently lora adapter is applied to layers in
llama_apply_lora_from_file_internal
but i want to maintain multiple adapters without fusing with base model (not apply on top of another). maybe route to best adapter + base combo bashed on prompt using some another router adapter or could be small bert that tags type of field prompt is related etc.Might sound stupid but i want to play around these ideas any guidance on how to achieve something like that would very helpful thank you
Beta Was this translation helpful? Give feedback.
All reactions