Custom modes: is Claude unable to take instructions? or what's going on here? #1567
Unanswered
mindplay-dk
asked this question in
3. Q&A / helpdesk
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
After a very frustrating experience with the VS Code Copilot agent, I decided to give Kilo a shot.
I will say first, the results here are way better than with the mysterious black box that is Copilot agents - and I am extremely pleased with the fact that Kilo lets you preview the full system prompt!
This is an extremely important feature - with Copilot, I had absolutely zero insight, not the faintest clue what it was doing with my custom mode prompt or why. With Kilo, this is super clear and transparent! 🙌
With that said, and being able to see the system prompt now (which appears to make sense) I am now running into what appears to be an issue with Claude itself (?) where it simply does not follow instructions. 🤔
Minor issue, but the "export mode" button doesn't seem to do anything? Nothing appears where I ask it to save.
So here are the individual inputs instead:
Role Definition:
When to use:
Mode-specific custom instructions:
As you can see, I've been REALLY stern about trying to make it follow my prescribed workflow, and it just... doesn't. 🤷♂️
It insists on building out the whole feature in one shot, and then proceeds to debug - the point of this custom mode, as you can see, was to try to get it to follow my personal workflow, where I build and test smaller units in iterations.
It appears that's simply not something it wants to do? 🥲
But I'm not 100% clear on exactly what a Kilo agent does, so maybe someone can enlighten me... is it literally just sending the system prompt, as you can preview it? And then it passes the user's request and context? And then the LLM takes over from there, calling tools, etc. until the task is done?
Or is there more going on behind the scenes?
Because, if that's how it works, if it's literally asking the LLM to follow an entire process in a single round-trip, then, I'm starting to understand why I can't coerce it into following this workflow - it's simply not how it was trained to respond to coding tasks, is it?
I had kind of hoped agents were a little more than that?
I remember some of the very early Python-based CLI agents years ago, and how the first basic agents would always have two or more LLMs talking to each other. This seems to have fallen completely out of fashion, so I don't guess Kilo does anything like that?
I get that agents are now large and capable enough to solve tasks with a single LLM - and I'm sure this approach is overall faster, since you don't have to copy tokens back and forth, but... I wanted to test this iterative workflow, because, two reasons:
It doesn't seem all that efficient.
Buy anyhow, long story short, I guess I'm trying to learn whether something like this is even possible with modern agents?
Am I barking up entirely the wrong tree here? 😅
Because I imagine something like this would be possible if one LLM planned the task (in terms of the units to be implemented) and then passed the request to a second LLM for it to implement only that first unit, and a test. I would think that should be possible? Like, if the first LLM doesn't ask for a whole feature, there's no way the second LLM is going to try to build it, is there? Whereas, if you ask Claude to do something in a single turn, then yeah, it's going to do what it was trained to do.
Am I finally grasping this, or no? 🤷♂️😄
Any help and guidance would be greatly appreciated. 🙏
PS: Kilo built the same feature as Copilot with the same Claude 4 model on half the budget. Really impressive! 🙌
Beta Was this translation helpful? Give feedback.
All reactions