On-Device Zeta #24859
Replies: 11 comments 9 replies
-
I don't really disagree though this is an OS model (as is Zed itself) and Zed does interface with Ollama and LM Studio so maybe that's coming. Other than that, I would very much like to see Zed continue on its own. Presumably at a certain point they'll need to monetize somehow or the logical conclusion is getting acquired by a large company with varied motivations and priorities. |
Beta Was this translation helpful? Give feedback.
-
I'm going to chime in here and give some of my perspective. I respect Zed team's decision of releasing Zeta fully opensource & open-data, and I guess that on-device zeta is a direction that they will take in the near future. After all, Zed already allow fully customizable LLM source for the assistant, it would make no sense that they will paywall this feature. I would love to get the Zed team's input on this. |
Beta Was this translation helpful? Give feedback.
-
Similarly, it would be nice to have the inline edit predictions with other models like a fat Qwen 32b model. I can manually make calls myself with Ollama, but I want it integrated into my awesome editor |
Beta Was this translation helpful? Give feedback.
-
I've been working on a personal project called zedex to which I recently added a crude implementation of edit prediction for Zed using any OpenAI-compatible backend. Some of the bits in the implementation are just winged / didn't really look too close at the Zed implementation but it seems to work OK for the small things I use it for. At the moment is uses some serious prompt engineering (lol) as glue to work well with both Llama 3.3 70B and the newer Llama 4 models. I haven't tried the Zeta fined-tuned model. At the moment it requires you to give up a lot of other Zed features including (and not limited to) login and collaboration as you have to override |
Beta Was this translation helpful? Give feedback.
-
would be nice to get it to work locally. Recently it seem like I hit a limit of free predictions on my free plan and was wondering what's next. I already run qwen2.5 coder locally, and since zeta also based on same model, I think it would be nice to run it all locally |
Beta Was this translation helpful? Give feedback.
-
this is a no brainer, people are not paying 20 a month to get edit predictions from qwen2.5coder 7b, an open source, small and limited model, theyre paying for api access to the latest big time models with agenetic editing zeta is open source, the community contributed to it, and any half decent hardware can run it locally, whatever it costs to run edit predictions for zed would be practically wiped out windsurf has separated edit prediction from the main subscription, its unlimited on free tier and its a better model/ux than zeta, although not as good as Cursor Tab, zeta is in 3rd place at the minute for edit prediction and theres just no justification for locking it behind a 20 a month subscription the people who want chat, agents and collab are going to subscribe regardless of whether zeta can be run locally or not, and people who only use edit prediction are not going to subscribe when better is available elsewhere for free |
Beta Was this translation helpful? Give feedback.
-
I would pay for Zed to get this feature, so to me it's not a $$$ question it's a flexibility and capability question. I'd like to be able to set up and configure arbitrary local or remote providers as edit predictors with a lot of flexibility. |
Beta Was this translation helpful? Give feedback.
-
+1 would also really appreciate this feature |
Beta Was this translation helpful? Give feedback.
-
I would really like to hack on this feature, try other models, see the difference, try their fine tuning and improve on it. Not being able to execute this locally - at least I couldn't yet find a way - makes this really difficult. |
Beta Was this translation helpful? Give feedback.
-
This would be a much appreciated feature. Vscode allows code autocomplete via ollama with the continue plugin. |
Beta Was this translation helpful? Give feedback.
-
I might pay to have Gemini do my code predictions, not being able to do it is highly motivating to refuse to pay a subscription and move away from zed. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Putting Zeta on servers and relying on data center compute, plus network transfer speed, is expensive in both latency and literal server cost. Why not, for the hardware that can reasonably support it, put Zeta on-device? This would potentially allow for more reliable latency and offline use, in addition to being better for privacy.
Either that, or an option to do completions on-device through Ollama.
It's baffling to me that companies continue to not utilize local compute. I don't want to pay yet another subscription to access tools like this. I paid for a powerful local system - use it!
Beta Was this translation helpful? Give feedback.
All reactions