Jochen
fb2cf6f898
fix for thinking budget and hardcoded max token on summary
2025-12-09 12:41:52 +11:00
Jochen
696c441a47
validate max_tokens for call, also fallbacks for summary
...
When the CW is full, max_tokens is often passed at 0 or tiny. The LLM will fail. For Anthropic with thining, there is also the thinking budget.
This can happen during summary attempts, in that case
first try thinnify, skinnify etc..
2025-12-09 10:15:32 +11:00
Dhanji R. Prasanna
48e6d594bc
tweak todo tool output
2025-12-08 11:05:01 +11:00
Dhanji R. Prasanna
678403da35
add a force thinnify cmd
2025-12-05 15:32:13 +11:00
Jochen
928f2bfa9d
actually record coach feedback and use it
2025-12-02 21:23:50 +11:00
Jochen
21af6ba574
fix temperature for summary request too.
2025-12-02 21:20:16 +11:00
Jochen
ae16243f49
Fix temperature param + add thinking for anthropic
...
The temperature param was not passed to the llm.
Now support anthropic models in 'thinking' mode.
2025-12-02 17:24:55 +11:00
Dhanji R. Prasanna
d9ad244197
add markdown format only to final_output and fix todo duplication
2025-12-02 14:26:22 +11:00
Dhanji R. Prasanna
a6537e4dba
todo_write outputs entire list
2025-12-02 13:48:05 +11:00
Dhanji R. Prasanna
f8f989d4c6
resume unfinished TODOs
2025-12-02 11:06:58 +11:00
Dhanji R. Prasanna
1b4ea93ba4
token counting bugfix
2025-12-01 14:52:10 +11:00
Dhanji R. Prasanna
4496eee046
fix compaction to restore system message
2025-12-01 14:38:21 +11:00
Dhanji R. Prasanna
81fd2ab92f
unused var
2025-11-29 15:44:30 +11:00
Jochen
dcfd681b05
add summary context window
2025-11-28 16:33:31 +11:00
Jochen
52f78653b4
add context window monitor
...
Writes the current context window to logs/current_context_window (uses a symlink to a session ID).
This PR was unfortunately generated by a different LLM and did a ton of superficial reformating, it's actually a fairly small and benign change, but I don't want to roll back everything. Hope that's ok.
2025-11-27 21:00:02 +11:00
Jochen
5170744099
add cache_control to user messages
2025-11-27 13:12:42 +11:00
Jochen
c837308148
never add more than 4 cache controls
...
Anthropic API throws errors otherwise.
2025-11-26 18:38:30 +11:00
Jochen
1e1702001c
Add logging for discovery
2025-11-26 10:41:35 +11:00
Jochen
ad198a8501
add code exploration fast start
...
This tries to short-circuit multiple round-trips to llm for reading code.
It's a precursor to trying to context engineer tailored to specific tasks.
In initial experiments, it's only marginally faster than regular mode, and burns more tokens.
2025-11-25 22:51:32 +11:00
Jochen
a96a15d1fc
add code coverage command
2025-11-21 14:38:58 +11:00
Jochen
551a577ee1
changed user choice for TODO stale check
...
user can ignore, mark stale or quit.
2025-11-21 12:35:14 +11:00
Jochen
84718223bc
remove minor comment
2025-11-21 12:26:41 +11:00
Jochen
28a83d2dcf
check for stale TODOs
...
on by default, can be disabled
2025-11-21 12:09:01 +11:00
Jochen
be6c6bfca4
fix ref to system prompt
2025-11-21 10:49:39 +11:00
Jochen
94a41c5c34
don't write warning to console
2025-11-21 10:49:27 +11:00
Jochen
09dbad2d68
allow multiple tool calls, log warnings if there are duplicate calls.
...
controlled via a flag to the agent config:
allow_multiple_tool_calls = true
2025-11-21 10:49:15 +11:00
Jochen
ffbf410b17
log tool calls
2025-11-21 10:49:02 +11:00
Dhanji Prasanna
14c8d066c9
ensure system prompt is always added first
2025-11-20 08:45:03 +11:00
Jochen
b6e226df67
Merge pull request #23 from dhanji/jochen-add-code-instructions
...
system prompt now includes code style guide
2025-11-19 16:25:20 +11:00
Jochen
1069664e16
fix bad max_tokens and context_window logic
...
for non-databricks code
2025-11-19 13:51:16 +11:00
Jochen
9bffd8b1bf
cache_control removed from databricks
2025-11-19 12:15:49 +11:00
Jochen
bfee8040e9
regression tests added
2025-11-19 11:32:14 +11:00
Jochen
a150ba6a55
adds ttl to cache control
2025-11-18 23:23:49 +11:00
Jochen
296bf5a449
adds cache_control
2025-11-18 22:38:52 +11:00
Jochen
7f73b664a3
system prompt now includes code style guide
2025-11-18 18:21:16 +11:00
Michael Neale
81cd956c20
allow openai to be used to name named compatible providers
2025-11-10 16:12:33 +11:00
Jochen
dce0d08f8c
fix OpenAI max_token config read
2025-11-10 11:58:34 +11:00
Dhanji Prasanna
1f12ff6ca0
fix refresh and max_tokens bug
2025-11-07 09:50:43 +11:00
Dhanji Prasanna
cb43fcdecf
g3 console init
2025-11-07 09:29:29 +11:00
Dhanji Prasanna
aaf918828f
g3 console initial cut + error doesnt kill auto
2025-11-07 09:27:13 +11:00
Dhanji R. Prasanna
6913c5f72e
Merge pull request #19 from dhanji/jochen-fix-anthropic-context
...
Fix context window exhaustion
2025-11-07 08:29:01 +11:00
Jochen
0e1f9dbf9a
rename max_context_length to fallback_default_max_tokens
2025-11-06 19:47:02 +11:00
Dhanji R. Prasanna
8eda691cb1
todo persistence
2025-11-06 15:24:57 +11:00
Jochen
af20c93c61
respect context length for anthropic
...
use the context length as per the config, rather than just hard-coded values.
2025-11-06 15:07:46 +11:00
Dhanji R. Prasanna
f61b0d000c
small todo fix
2025-11-06 14:53:06 +11:00
Dhanji R. Prasanna
624ca65e2e
encourage use of todo tools
2025-11-06 14:30:00 +11:00
Dhanji R. Prasanna
d78732df14
colors
2025-11-06 13:41:06 +11:00
Dhanji R. Prasanna
d007e8f471
improve code_search nudge and increase anthropic tmieout
2025-11-05 15:05:29 +11:00
Dhanji R. Prasanna
4327c839a9
added scheme and kotlin to code_search
2025-11-05 14:17:15 +11:00
Dhanji R. Prasanna
fa38439a06
adding more languages to tree-sitter (java, go, cpp,..)
2025-11-05 14:07:50 +11:00