A persistent issue I have with Cursor et al. is that they hallucinate function arguments when using a function or method from a library. It seems like automatically pulling the library's documentation into the context would be helpful, but I haven't found any tool that does this automatically. Is there any chance that Runner does this?
It doesn't seem like this was the problem you were trying to solve, but reliable use of libraries and APIs is a critical problem to solve if you want LLM-generated code to work.
I’ve largely solved this with the context7 mcp server. Any time my prompt is likely to touch apis I know the LLM will get wrong I tell it to review the docs with context7 first.
I agree that's a major problem. It's not something I've solved yet. I suspect a web research sub-agent is likely what's needed, so it can pull in up-to-date docs for whatever library you need to work with.
I think there's probably a lot of value to be gained in tooling for coding agents that codify and enhance the describe -> explore -> plan -> refine -> implement -> verify cycle. With most popular tools (cursor, claude code, roo, augment, windsurf, etc) you have to do this workflow "manually" usually by having the model write out .md files, it isn't super smooth.
Cool, I hadn't heard of Traycer. That does look quite similar!
Completely agree. I basically built Runner to codify the process I was already using manually with Claude Code and Gemini. A lot of developers seem to be settling on a similar workflow, so I'm betting that something like Runner or Traycer will be useful for a lot of devs.
I'll be curious to see how far the existing players like Cursor push in this direction.
Gemini is required for the context management sub-agent. You can use any of OpenAI, Anthropic, or Gemini for the main planning and coding agents, but GPT-5 performs the best in my experience. Claude 4 Sonnet works well too, but it's about twice as expensive.
Interesting, will check it out.
A persistent issue I have with Cursor et al. is that they hallucinate function arguments when using a function or method from a library. It seems like automatically pulling the library's documentation into the context would be helpful, but I haven't found any tool that does this automatically. Is there any chance that Runner does this?
It doesn't seem like this was the problem you were trying to solve, but reliable use of libraries and APIs is a critical problem to solve if you want LLM-generated code to work.
I’ve largely solved this with the context7 mcp server. Any time my prompt is likely to touch apis I know the LLM will get wrong I tell it to review the docs with context7 first.
Good to know. I've heard great things about Context7, but haven't experimented with it yet.
I agree that's a major problem. It's not something I've solved yet. I suspect a web research sub-agent is likely what's needed, so it can pull in up-to-date docs for whatever library you need to work with.
Cool stuff! This looks quite similar to https://traycer.ai/
I think there's probably a lot of value to be gained in tooling for coding agents that codify and enhance the describe -> explore -> plan -> refine -> implement -> verify cycle. With most popular tools (cursor, claude code, roo, augment, windsurf, etc) you have to do this workflow "manually" usually by having the model write out .md files, it isn't super smooth.
Cool, I hadn't heard of Traycer. That does look quite similar!
Completely agree. I basically built Runner to codify the process I was already using manually with Claude Code and Gemini. A lot of developers seem to be settling on a similar workflow, so I'm betting that something like Runner or Traycer will be useful for a lot of devs.
I'll be curious to see how far the existing players like Cursor push in this direction.
Tried downloading it from my M2 macbook pro 2 times and failed
As in the download itself didn't happen when you clicked the download button, or the installation failed?
> You will need at least a Gemini API key, and for best performance also an OpenAI API key.
Could you explain why both / why not also Claude (why not all three?)
Gemini is required for the context management sub-agent. You can use any of OpenAI, Anthropic, or Gemini for the main planning and coding agents, but GPT-5 performs the best in my experience. Claude 4 Sonnet works well too, but it's about twice as expensive.