Replies: 1 comment 2 replies
-
I have a command that works well for speculative decoding on my system - Now, the question is, how can I offload the draft model to my other mac mini (M2)? I have doubts if this would end up benefitting me (I guess the draft model needs to speak with the main model quite frequently, and latency should be important; I'm not sure we get it with Ethernet or Thunderbolt 4). But, as in the case of any experiment, trying it out, and seeing how bad/good it actually is, would be worth it right? I don't understand |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I've got two mac mins - one with 16GB RAM (M2 Pro), and the other with 8GB RAM (M2). Now, I was wondering if I can leverage the power of speculative decoding to speed up inference perforamance of a main model (like a Qwen2.5-Coder-14B 4bits quantized GGUF) on the M2 Pro mac, while having the draft model (like a Qwen2.5-Coder-0.5B 8bits quantized GGUF) running via the M2 mac. Is this feasible, perhaps using
rpc-server
? Can someone who's done something like this help me out please? Also, if this is possible, is it scalable even further (I have an old desktop with an RTX 2060).A related discussion, which I couldn't quite understand happened earlier.
Beta Was this translation helpful? Give feedback.
All reactions