# Workspace Memory > Updated: 2026-01-27T03:15:00Z | Size: ~16k chars ### Remember Tool Wiring - `crates/g3-core/src/tools/memory.rs` [0..5000] - `execute_remember()`, `get_memory_path()`, `merge_memory()` - `crates/g3-core/src/tool_definitions.rs` [11000..12000] - remember tool definition in `create_core_tools()` - `crates/g3-core/src/tool_dispatch.rs` [48] - dispatch case for "remember" - `crates/g3-core/src/prompts.rs` [4200..6500] - Workspace Memory section in native prompt - `crates/g3-cli/src/project_files.rs` - `read_workspace_memory()` loads from `analysis/memory.md` ### Context Window & Compaction Token tracking, history management, and context compaction. - `crates/g3-core/src/context_window.rs` [0..815] - `ContextWindow` - token tracking, message history - `reset_with_summary()` - compact history to summary - `should_compact()` - threshold check (80%) - `thin_context()` - replace large results with file refs - `ThinResult` [16..36] - struct with scope, before/after %, chars_saved - `crates/g3-core/src/compaction.rs` [0..11213] - `perform_compaction()` - unified compaction for force_compact() and auto-compaction - `CompactionResult`, `CompactionConfig` - `calculate_capped_summary_tokens()`, `should_disable_thinking()` - `build_summary_messages()`, `apply_summary_fallback_sequence()` - `crates/g3-core/src/lib.rs` - `Agent.force_compact()`, `stream_completion_with_tools()` ### Session Storage & Continuation - `crates/g3-core/src/session_continuation.rs` [0..541] - `SessionContinuation`, `save_continuation()`, `load_continuation()` - `crates/g3-core/src/paths.rs` [0..133] - `get_session_logs_dir()`, `get_thinned_dir()`, `get_session_file()` - `crates/g3-core/src/session.rs` - Session logging utilities ### Tool System - `crates/g3-core/src/tool_definitions.rs` [0..544] - `create_core_tools()`, `create_tool_definitions()`, `ToolConfig` - `crates/g3-core/src/tool_dispatch.rs` [0..73] - `dispatch_tool()` routing ### CLI Module Structure Refactored CLI with extracted modules for each mode. - `crates/g3-cli/src/lib.rs` [0..415] - Entry point, `run()`, mode dispatch, config loading - `crates/g3-cli/src/cli_args.rs` [0..133] - `Cli` struct with clap derive macros - `crates/g3-cli/src/autonomous.rs` [0..785] - `run_autonomous()`, coach-player feedback loop - `crates/g3-cli/src/agent_mode.rs` [0..284] - `run_agent_mode()`, `Agent::new_with_custom_prompt()` - `crates/g3-cli/src/accumulative.rs` [0..343] - `run_accumulative_mode()` iterative requirements - `crates/g3-cli/src/interactive.rs` [0..851] - `run_interactive()`, `run_interactive_machine()`, REPL - `crates/g3-cli/src/task_execution.rs` [0..212] - `execute_task_with_retry()`, `OutputMode` enum - `crates/g3-cli/src/commands.rs` [17..320] - `/help`, `/compact`, `/thinnify`, `/fragments`, `/rehydrate`, etc. - `crates/g3-cli/src/utils.rs` [0..91] - `display_welcome_message()`, `get_workspace_path()` - `crates/g3-cli/src/display.rs` - `format_workspace_path()`, `LoadedContent`, `print_loaded_status()` ### Auto-Memory System Auto-prompts LLM to save discoveries after tool calls. - `crates/g3-core/src/lib.rs` - `send_auto_memory_reminder()` [47800..48800] - MEMORY CHECKPOINT prompt with few-shot examples - `set_auto_memory()` [1451..1454] - enable/disable - `tool_calls_this_turn: Vec` [116] - tracks tools per turn - `execute_tool_in_dir()` [2843..2855] - records tool calls - `crates/g3-core/src/prompts.rs` [3800..4500] - Memory Format section in system prompt - `crates/g3-cli/src/lib.rs` [393] - `--auto-memory` CLI flag ### Streaming Markdown Formatter Terminal markdown rendering with syntax highlighting. - `crates/g3-cli/src/streaming_markdown.rs` - `format_header()` [21500..22500] - headers with inline formatting - `process_in_code_block()` [439..462] - detects closing fence - `emit_code_block()` [654..675] - joins buffer, highlights code - `flush_incomplete()` [693..735] - handles unclosed blocks at stream end - `crates/g3-cli/tests/streaming_markdown_test.rs` - `test_bold_inside_header`, `test_code_inside_header`, etc. - **Gotcha**: closing ``` without trailing newline must be detected in `flush_incomplete()` ### Retry Infrastructure - `crates/g3-core/src/retry.rs` [0..12000] - `execute_with_retry()`, `retry_operation()`, `RetryConfig`, `RetryResult` - Used by g3-planner; g3-cli has `execute_task_with_retry()` in task_execution.rs ### UI Abstraction Layer - `crates/g3-core/src/ui_writer.rs` [0..4500] - `UiWriter` trait, `NullUiWriter`, `print_thin_result()` - `crates/g3-cli/src/ui_writer_impl.rs` [0..14000] - `ConsoleUiWriter`, `print_tool_compact()` - `crates/g3-cli/src/simple_output.rs` [0..1200] - `SimpleOutput` helper ### Feedback Extraction - `crates/g3-core/src/feedback_extraction.rs` [0..22000] - `extract_coach_feedback()`, `try_extract_from_session_log()`, `try_extract_from_native_tool_call()` - `crates/g3-cli/src/coach_feedback.rs` [0..4025] - `extract_from_logs()` for coach-player loop ### Streaming Utilities & State - `crates/g3-core/src/streaming.rs` - `truncate_line()`, `truncate_for_display()`, `log_stream_error()`, `is_connection_error()` - `StreamingState` [17..42] - cross-iteration: full_response, first_token_time, iteration_count - `IterationState` [65..90] - per-iteration: parser, current_response, tool_executed - `MAX_ITERATIONS` [15] - constant (400) - `ToolOutputFormat` [100..112] - enum: SelfHandled, Compact(String), Regular - `format_tool_result_summary()`, `is_compact_tool()`, `format_compact_tool_summary()` - `crates/g3-core/src/lib.rs` [1879..2712] - `stream_completion_with_tools()` main loop (834 lines) ### Background Process Management - `crates/g3-core/src/background_process.rs` [0..3000] - `BackgroundProcessManager`, `start()`, `list()`, `is_running()`, `get()`, `remove()` - No `stop()` method - use shell tool with `kill ` ### Unified Diff Application - `crates/g3-core/src/utils.rs` [5000..15000] - `apply_unified_diff_to_string()`, `parse_unified_diff_hunks()` - Handles multi-hunk diffs, CRLF normalization, range constraints ### Error Classification - `crates/g3-core/src/error_handling.rs` [0..567] - `classify_error()`, `ErrorType`, `RecoverableError` - Priority: rate limit > network > server > busy > timeout > token limit > context length - **Gotcha**: "Connection timeout" → NetworkError (not Timeout) due to "connection" keyword priority ### CLI Metrics - `crates/g3-cli/src/metrics.rs` [0..5416] - `TurnMetrics`, `format_elapsed_time()`, `generate_turn_histogram()` ### ACD (Aggressive Context Dehydration) Saves conversation fragments to disk, replaces with stubs. - `crates/g3-core/src/acd.rs` [0..22000] - `Fragment` - `new()`, `save()`, `load()`, `generate_stub()`, `list_fragments()`, `get_latest_fragment_id()` - `crates/g3-core/src/tools/acd.rs` [0..8500] - `execute_rehydrate()` tool - `crates/g3-core/src/paths.rs` [3200..3400] - `get_fragments_dir()` → `.g3/sessions//fragments/` - `crates/g3-core/src/compaction.rs` [195..240] - ACD integration, creates fragment+stub when enabled - `crates/g3-core/src/context_window.rs` [10100..10700] - `reset_with_summary_and_stub()` - `crates/g3-cli/src/lib.rs` [157..161] - `--acd` flag; [1476..1525] - `/fragments`, `/rehydrate` commands **Fragment JSON fields**: `fragment_id`, `created_at`, `messages`, `message_count`, `user_message_count`, `assistant_message_count`, `tool_call_summary`, `estimated_tokens`, `topics`, `preceding_fragment_id` ### UTF-8 Safe String Slicing Rust `&s[..n]` uses byte indices; multi-byte chars (emoji, CJK) panic if sliced mid-character. **Pattern**: `s.char_indices().nth(n).map(|(i,_)| i).unwrap_or(s.len())` for byte index of Nth char. **Danger zones**: Display truncation, ACD stubs, user input, any non-ASCII text. ### Studio - Multi-Agent Workspace Manager - `crates/studio/src/main.rs` [0..12500] - `cmd_run()`, `cmd_status()`, `cmd_accept()`, `cmd_discard()`, `extract_session_summary()` - `crates/studio/src/session.rs` - `Session`, `SessionStatus` - `crates/studio/src/git.rs` - `GitWorktree` for isolated agent sessions **Session log path**: `/.g3/sessions//session.json` **Fields**: `context_window.{conversation_history, percentage_used, total_tokens, used_tokens}`, `session_id`, `status`, `timestamp` ### Racket Code Search Support - `crates/g3-core/src/code_search/searcher.rs` - Racket parser [~line 45] - `tree_sitter_racket::LANGUAGE` - Extensions [~line 90] - `.rkt`, `.rktl`, `.rktd` → "racket" ### Language-Specific Prompt Injection Auto-detects languages and injects toolchain guidance. - `crates/g3-cli/src/language_prompts.rs` - `LANGUAGE_PROMPTS` [12..19] - (lang_name, extensions, prompt_content) array - `detect_languages()` [22..32] - scans workspace - `get_language_prompts_for_workspace()` [88..108] - returns formatted prompt - `scan_directory_for_extensions()` [42..77] - recursive, depth limit 2, skips hidden/vendor - `AGENT_LANGUAGE_PROMPTS` [21..26] - (agent_name, lang_name, prompt_content) - `get_agent_language_prompts_for_workspace()` [124..137] - agent+lang lookup - `crates/g3-cli/src/agent_mode.rs` [149..159] - appends agent-specific prompts - `prompts/langs/` - language prompt files (e.g., `racket.md`, `carmack.racket.md`) **To add language**: Create `prompts/langs/.md`, add to `LANGUAGE_PROMPTS` **To add agent+lang**: Create `prompts/langs/..md`, add to `AGENT_LANGUAGE_PROMPTS` ### MockProvider for Testing - `crates/g3-providers/src/mock.rs` - `MockProvider` [220..320] - response queue, request tracking - `MockResponse` [35..200] - configurable chunks and usage - `scenarios` module [410..480] - `text_only_response()`, `multi_turn()`, `tool_then_response()` - `crates/g3-core/tests/mock_provider_integration_test.rs` - `test_butler_bug_scenario()`, `test_multi_turn_text_only_maintains_alternation()` **Usage**: `MockProvider::new().with_response(MockResponse::text("Hello!"))` ### G3 Status Message Formatting - `crates/g3-cli/src/g3_status.rs` - `G3Status` - static methods for "g3:" prefixed messages - `Status` enum - `Done`, `Failed`, `Error(String)`, `Custom(String)`, `Resolved`, `Insufficient`, `NoChanges` - `progress()` [64..76] - "g3: ..." (no newline) - `done()` [93..101] - bold green "[done]" - `failed()` [104..111] - red "[failed]" - `thin_result()` [265..292] - formats ThinResult with colors - `resuming()` [227..236] - session resume with cyan ID ### Prompt Cache Statistics Tracks prompt/prefix caching across providers. - `crates/g3-providers/src/lib.rs` [195..210] - `Usage.cache_creation_tokens`, `cache_read_tokens` - `crates/g3-providers/src/anthropic.rs` [944..956] - parses `cache_creation_input_tokens`, `cache_read_input_tokens` - `crates/g3-providers/src/openai.rs` [494..510] - parses `prompt_tokens_details.cached_tokens` - `crates/g3-core/src/lib.rs` [75..90] - `CacheStats` struct; [106] - `Agent.cache_stats` - `crates/g3-core/src/stats.rs` [189..230] - `format_cache_stats()` with hit rate metrics ### Embedded Provider (Local LLM) Local inference via llama-cpp-rs with Metal acceleration. - `crates/g3-providers/src/embedded.rs` - `EmbeddedProvider` [22..85] - session, model_name, max_tokens, temperature, context_length - `new()` [26..85] - tilde expansion, auto-downloads Qwen if missing - `format_messages()` [87..175] - converts to prompt string (Qwen/Mistral/Llama templates) - `get_stop_sequences()` [280..340] - model-specific stop tokens - `stream()` [560..780] - via spawn_blocking + mpsc **Known issues**: Provider name hardcoded as "embedded"; missing GLM-4 template; no streaming usage tracking. ### Chat Template Formats | Model | Format | Start Token | End Token | |-------|--------|-------------|----------| | Qwen | ChatML | `<\|im_start\|>role\n` | `<\|im_end\|>` | | GLM-4 | ChatGLM4 | `[gMASK]<\|role\|>\n` | `<\|endoftext\|>` | | Mistral | Instruct | `[INST]` | `[/INST]` | | Llama | Llama2 | `<>` | `<>` | ### Recommended GGUF Models | Model | Size | Use Case | |-------|------|----------| | GLM-4-9B-Q8_0 | ~10GB | Fast, capable | | GLM-4-32B-Q6_K_L | ~27GB | Top tier coding/reasoning | | Qwen3-4B-Q4_K_M | ~2.3GB | Small, rivals 72B | **Download**: `python3 -m huggingface_hub.commands.huggingface_cli download --include "" --local-dir ~/.g3/models/` **Config**: ```toml [providers.embedded.glm4] model_path = "~/.g3/models/THUDM_GLM-4-32B-0414-Q6_K_L.gguf" model_type = "glm4" context_length = 32768 max_tokens = 4096 gpu_layers = 99 ```