G3 is a coding AI agent designed to help you complete tasks by writing code and executing commands. Built in Rust, it provides a flexible architecture for interacting with various Large Language Model (LLM) providers while offering powerful code generation and task automation capabilities.
G3 follows a modular architecture organized as a Rust workspace with multiple crates, each responsible for specific functionality:
The heart of the agent system, containing:
- Agent Engine: Main orchestration logic for handling conversations, tool execution, and task management
- Context Window Management: Intelligent tracking of token usage with context thinning (50-80%) and auto-summarization at 80% capacity
- Tool System: Built-in tools for file operations, shell commands, computer control, TODO management, and structured output
- Streaming Response Parser: Real-time parsing of LLM responses with tool call detection and execution
- Task Execution: Support for single and iterative task execution with automatic retry logic
Abstraction layer for LLM providers:
- Provider Interface: Common trait-based API for different LLM backends
- Multiple Provider Support:
- Anthropic (Claude models)
- Databricks (DBRX and other models)
- Local/embedded models via llama.cpp with Metal acceleration on macOS
- OAuth Authentication: Built-in OAuth flow support for secure provider authentication
- Provider Registry: Dynamic provider management and selection
Configuration management system:
- Environment-based configuration
- Provider credentials and settings
- Model selection and parameters
- Runtime configuration options
Task execution framework:
- Task planning and decomposition
- Execution strategies (sequential, parallel)
- Error handling and retry mechanisms
- Progress tracking and reporting
Computer control capabilities:
- Mouse and keyboard automation
- UI element inspection and interaction
- Screenshot capture and window management
- OCR text extraction via Tesseract
Command-line interface:
- Interactive terminal interface
- Task submission and monitoring
- Configuration management commands
- Session management
G3 includes robust error handling with automatic retry logic:
- Recoverable Error Detection: Automatically identifies recoverable errors (rate limits, network issues, server errors, timeouts)
- Exponential Backoff with Jitter: Implements intelligent retry delays to avoid overwhelming services
- Detailed Error Logging: Captures comprehensive error context including stack traces, request/response data, and session information
- Error Persistence: Saves detailed error logs to
logs/errors/for post-mortem analysis - Graceful Degradation: Non-recoverable errors are logged with full context before terminating
G3 includes intelligent duplicate detection to prevent the LLM from accidentally calling the same tool twice in a row:
- Sequential Duplicate Prevention: Only immediately sequential identical tool calls are blocked
- Text Separation Allowed: If there's any text between tool calls, they're not considered duplicates
- Session-Wide Reuse: Tools can be called multiple times throughout a session - only back-to-back duplicates are prevented
This catches cases where the LLM "stutters" and outputs the same tool call twice, while still allowing legitimate re-use of tools.
After each response, G3 displays a timing footer showing elapsed time, time to first token, token usage (from the LLM, not estimated), and current context window usage percentage. The token and context info is displayed dimmed for a clean interface.
- Automatic context window monitoring with percentage-based tracking
- Smart auto-summarization when approaching token limits
- Context thinning at 50%, 60%, 70%, 80% thresholds - automatically replaces large tool results with file references
- Conversation history preservation through summaries
- Dynamic token allocation for different providers (4k to 200k+ tokens)
G3's interactive CLI includes control commands for manual context management:
/compact: Manually trigger summarization to compact conversation history/thinnify: Manually trigger context thinning to replace large tool results with file references/skinnify: Manually trigger full context thinning (like/thinnifybut processes the entire context window, not just the first third)/readme: Reload README.md and AGENTS.md from disk without restarting/stats: Show detailed context and performance statistics/help: Display all available control commands
These commands give you fine-grained control over context management, allowing you to proactively optimize token usage and refresh project documentation. See Control Commands Documentation for detailed usage.
- File Operations: Read, write, and edit files with line-range precision
- Shell Integration: Execute system commands with output capture
- Code Generation: Structured code generation with syntax awareness
- TODO Management: Read and write TODO lists with markdown checkbox format
- Computer Control (Experimental): Automate desktop applications
- Mouse and keyboard control
- macOS Accessibility API for native app automation (via
--macaxflag) - UI element inspection
- Screenshot capture and window management
- OCR text extraction from images and screen regions
- Window listing and identification
- Code Search: Embedded tree-sitter for syntax-aware code search (Rust, Python, JavaScript, TypeScript, Go, Java, C, C++) - see Code Search Guide
- Final Output: Formatted result presentation
- Flock Mode: Parallel multi-agent development for large projects - see Flock Mode Guide
- Support for multiple LLM providers through a unified interface
- Hot-swappable providers without code changes
- Provider-specific optimizations and feature support
- Local model support for offline operation
- Single-shot task execution for quick operations
- Iterative task mode for complex, multi-step workflows
- Automatic error recovery and retry logic
- Progress tracking and intermediate result handling
- Language: Rust (2021 edition)
- Async Runtime: Tokio for concurrent operations
- HTTP Client: Reqwest for API communications
- Serialization: Serde for JSON handling
- CLI Framework: Clap for command-line parsing
- Logging: Tracing for structured logging (INFO logs converted to DEBUG for cleaner CLI output)
- Local Models: llama.cpp with Metal acceleration support
G3 is designed for:
- Automated code generation and refactoring
- File manipulation and project scaffolding
- System administration tasks
- Data processing and transformation
- API integration and testing
- Documentation generation
- Complex multi-step workflows
- Parallel development of modular architectures
- Desktop application automation and testing
The default interactive mode now uses accumulative autonomous mode, which combines the best of interactive and autonomous workflows:
# Simply run g3 in any directory
g3
# You'll be prompted to describe what you want to build
# Each input you provide:
# 1. Gets added to accumulated requirements
# 2. Automatically triggers autonomous mode (coach-player loop)
# 3. Implements your requirements iteratively
# Example session:
requirement> create a simple web server in Python with Flask
# ... autonomous mode runs and implements it ...
requirement> add a /health endpoint that returns JSON
# ... autonomous mode runs again with both requirements ...# Single-shot mode (one task, then exit)
g3 "implement a function to calculate fibonacci numbers"
# Traditional autonomous mode (reads requirements.md)
g3 --autonomous
# Traditional chat mode (simple interactive chat without autonomous runs)
g3 --chatPlanning mode provides a structured workflow for requirements-driven development with git integration:
# Start planning mode for a codebase
g3 --planning --codepath ~/my-project --workspace ~/g3_workspace
# Without git operations (for repos not yet initialized)
g3 --planning --codepath ~/my-project --no-git --workspace ~/g3_workspacePlanning mode workflow:
- Refine Requirements: Write requirements in
<codepath>/g3-plan/new_requirements.md, then let the LLM suggest improvements - Implement: Once requirements are approved, they're renamed to
current_requirements.mdand the coach/player loop implements them - Complete: After implementation, files are archived with timestamps (e.g.,
completed_requirements_2025-01-15_10-30-00.md) - Git Commit: Staged files are committed with an LLM-generated commit message
- Repeat: Return to step 1 for the next iteration
All planning artifacts are stored in <codepath>/g3-plan/:
planner_history.txt- Audit log of all planning activitiesnew_requirements.md/current_requirements.md- Active requirementstodo.g3.md- Implementation TODO listcompleted_*.md- Archived requirements and todos
See the configuration section for setting up different providers for the planner role.
# Build the project
cargo build --release
# Run from the build directory
./target/release/g3
# Or copy both files to somewhere in your PATH (macOS only needs both files)
cp target/release/g3 ~/.local/bin/
cp target/release/libVisionBridge.dylib ~/.local/bin/ # macOS only
# Execute a task
g3 "implement a function to calculate fibonacci numbers"G3 uses a TOML configuration file for settings. The config file is automatically created at ~/.config/g3/config.toml on first run with sensible defaults.
G3 includes configurable retry logic for handling recoverable errors (timeouts, rate limits, network issues, server errors):
[agent]
max_context_length = 8192
enable_streaming = true
timeout_seconds = 60
# Retry configuration for recoverable errors
max_retry_attempts = 3 # Default mode retry attempts
autonomous_max_retry_attempts = 6 # Autonomous mode retry attemptsRetry Behavior:
- Default Mode (
max_retry_attempts): Used for interactive chat and single-shot tasks. Default: 3 attempts. - Autonomous Mode (
autonomous_max_retry_attempts): Used for long-running autonomous tasks. Default: 6 attempts. - Retries use exponential backoff with jitter to avoid overwhelming services
- Autonomous mode spreads retries over ~10 minutes to handle extended outages
- Only recoverable errors are retried (timeouts, rate limits, 5xx errors, network issues)
- Non-recoverable errors (auth failures, invalid requests) fail immediately
Example: To increase timeout resilience in autonomous mode, set autonomous_max_retry_attempts = 10 in your config.
See config.example.toml for a complete configuration example.
G3 includes WebDriver support for browser automation tasks. Safari is the default, with Chrome headless available as an alternative.
One-Time Setup (macOS only):
Safari Remote Automation must be enabled before using WebDriver tools. Run this once:
# Option 1: Use the provided script
./scripts/enable-safari-automation.sh
# Option 2: Enable manually
safaridriver --enable # Requires password
# Option 3: Enable via Safari UI
# Safari → Preferences → Advanced → Show Develop menu
# Then: Develop → Allow Remote AutomationUsage:
# Use Safari (default, opens a visible browser window)
g3 --webdriver
# Use Chrome in headless mode (no visible window, runs in background)
g3 --chrome-headlessChrome Setup Options:
Option 1: Use Chrome for Testing (Recommended) - Guarantees version compatibility:
./scripts/setup-chrome-for-testing.shThen add to your ~/.config/g3/config.toml:
[webdriver]
chrome_binary = "/Users/yourname/.chrome-for-testing/chrome-mac-arm64/Google Chrome for Testing.app/Contents/MacOS/Google Chrome for Testing"Option 2: Use system Chrome - Requires matching ChromeDriver version:
- macOS:
brew install chromedriver - Linux:
apt install chromium-chromedriver - Or download from: https://chromedriver.chromium.org/downloads
Note: If you see "ChromeDriver version doesn't match Chrome version" errors, use Option 1 (Chrome for Testing) which bundles matching versions.
G3 includes support for controlling macOS applications via the Accessibility API, allowing you to automate native macOS apps.
Available Tools: macax_list_apps, macax_get_frontmost_app, macax_activate_app, macax_get_ui_tree, macax_find_elements, macax_click, macax_set_value, macax_get_value, macax_press_key
Setup: Enable with the --macax flag or in config with macax.enabled = true. Grant accessibility permissions:
- macOS: System Preferences → Security & Privacy → Privacy → Accessibility → Add your terminal app
For detailed documentation, see macOS Accessibility Tools Guide.
Note: This is particularly useful for testing and automating apps you're building with G3, as you can add accessibility identifiers to your UI elements.
G3 can interact with your computer's GUI for automation tasks:
Available Tools: mouse_click, type_text, find_element, take_screenshot, extract_text, find_text_on_screen, list_windows
Setup: Enable in config with computer_control.enabled = true and grant OS accessibility permissions:
- macOS: System Preferences → Security & Privacy → Accessibility
- Linux: Ensure X11 or Wayland access
- Windows: Run as administrator (first time only)
G3 automatically saves session logs for each interaction in the logs/ directory. These logs contain:
- Complete conversation history
- Token usage statistics
- Timestamps and session status
The logs/ directory is created automatically on first use and is excluded from version control.
MIT License - see LICENSE file for details
G3 is an open-source project. Contributions are welcome! Please see CONTRIBUTING.md for guidelines.