README
ΒΆ
AgentPipe π
AgentPipe is a powerful CLI and TUI application that orchestrates conversations between multiple AI agents. It allows different AI CLI tools (like Claude, Cursor, Gemini, Qwen, Ollama) to communicate with each other in a shared "room", creating dynamic multi-agent conversations with real-time metrics, cost tracking, and interactive user participation.
Screenshots
Enhanced TUI Interface
Enhanced TUI with multi-panel layout: agent list with status indicators, conversation view with metrics, statistics panel showing turns and total cost, configuration panel, and user input area
Supported AI Agents
- β Amp (Sourcegraph) - Advanced coding agent with autonomous reasoning
- β Claude (Anthropic) - Advanced reasoning and coding
- β Copilot (GitHub) - Terminal-based coding agent with multiple model support
- β Cursor (Cursor AI) - IDE-integrated AI assistance
- β Gemini (Google) - Multimodal understanding
- β Qwen (Alibaba) - Multilingual capabilities
- β Codex (OpenAI) - Code generation specialist
- β Ollama - Local LLM support
Features
Core Capabilities
- Multi-Agent Conversations: Connect multiple AI agents in a single conversation
- Multiple Conversation Modes:
round-robin: Agents take turns in a fixed orderreactive: Agents respond based on conversation dynamicsfree-form: Agents participate freely as they see fit
- Flexible Configuration: Use command-line flags or YAML configuration files
Enhanced TUI Interface
- Multi-panel layout with dedicated sections for agents, chat, stats, and config
- Color-coded agent messages with unique colors per agent
- Real-time agent activity indicators (π’ active/responding, β« idle)
- Inline metrics display (response time in seconds, token count, cost)
- Topic panel showing initial conversation prompt
- Statistics panel with turn counters and total conversation cost
- Configuration panel displaying all active settings and config file path
- Interactive user input panel for joining conversations
- Smart message consolidation (headers only on speaker change)
- Proper multi-paragraph message formatting
Production Features
- Prometheus Metrics: Comprehensive observability with 10+ metrics types
- Request rates, durations, errors
- Token usage and cost tracking
- Active conversations, retry attempts, rate limit hits
- HTTP server with
/metrics,/health, and web UI endpoints - Ready for Grafana dashboards and alerting
- Conversation Management:
- Save/resume conversations from state files
- Export to JSON, Markdown, or HTML formats
- Automatic chat logging to
~/.agentpipe/chats/
- Reliability & Performance:
- Rate limiting per agent with token bucket algorithm
- Retry logic with exponential backoff (configurable)
- Structured error handling with error types
- Config hot-reload for development workflows
- Middleware Pipeline: Extensible message processing
- Built-in: logging, metrics, validation, sanitization, filtering
- Custom middleware support for transforms and filters
- Error recovery and panic handling
- Docker Support: Multi-stage builds, docker-compose, production-ready
- Health Checks: Automatic agent health verification before conversations
- Agent Detection: Built-in doctor command to check installed AI CLIs
- Customizable Agents: Configure prompts, models, and behaviors for each agent
What's New π
Latest Updates (v0.0.16)
Major production-ready improvements since v0.0.15:
π Production-Ready Features
-
Prometheus Metrics: Comprehensive observability with 10+ metric types (1c3d3ac)
- HTTP server with
/metrics,/health, and web UI endpoints - Ready for Grafana dashboards and Prometheus alerting
- Track requests, durations, tokens, costs, errors, rate limits, retries
- OpenMetrics format support
- HTTP server with
-
Middleware Pipeline: Extensible message processing architecture (e44bcc6)
- 10+ built-in middleware (logging, metrics, validation, filtering, sanitization)
- Custom middleware support for transforms and filters
- Error recovery and panic handling
- Chain-of-responsibility pattern implementation
-
Conversation State Management: (cfde23b)
- Save/resume conversations from JSON state files
agentpipe resumecommand with --list flag- Automatic state directory (~/.agentpipe/states/)
- Full conversation history, config, and metadata preservation
-
Export Functionality: Multi-format conversation export (a4146ea)
- Export to JSON, Markdown, or HTML formats
- Professional HTML styling with responsive design
- XSS prevention with HTML escaping
agentpipe exportcommand
β‘ Reliability & Performance
-
Rate Limiting: Token bucket algorithm per agent (2ae8560)
- Configurable rate and burst capacity
- Thread-safe implementation with ~60ns overhead
- Automatic rate limit hit tracking in metrics
-
Retry Logic: Exponential backoff with smart defaults (ba16bf0)
- 3 retries with 1s initial delay, 30s max, 2.0x multiplier
- Configurable per orchestrator
- Retry attempt tracking in metrics
-
Structured Error Handling: Typed error system (ba16bf0)
- AgentError, ConfigError, ValidationError, TimeoutError, etc.
- Error wrapping with context
- Better error classification for metrics
-
Config Hot-Reload: Development workflow enhancement (abdab1c)
- Watch config files for changes with viper.WatchConfig
- Thread-safe reload with callbacks
- --watch-config flag for development mode
π³ Docker Support
Production-ready containerization (6cced13):
- Multi-stage Dockerfile (~50MB final image)
- docker-compose.yml with metrics server on port 9090
- Health checks and graceful shutdown
- Volume mounts for configs and logs
- Non-root user for security
- Complete documentation in docs/docker.md
π Testing & Quality
-
Comprehensive Test Coverage: 200+ tests (c0fb9c2, 39ede15, 62551c8)
- 86+ unit tests across orchestrator, adapters, logger, errors, ratelimit, config, conversation
- 15 integration tests for end-to-end conversation flows
- 25+ benchmark tests for performance regression detection
- TUI component tests with race detection
- All tests passing with concurrent access validation
-
Documentation: Complete docs/ directory (e562275)
- architecture.md - System design and patterns
- contributing.md - Contribution guidelines
- development.md - Development setup and workflows
- troubleshooting.md - Common issues and solutions
- docker.md - Docker deployment guide
π οΈ Developer Experience
-
Interactive Init Command: Configuration wizard (ba16bf0)
- Guided prompts for all configuration options
- Agent selection and configuration
- Orchestrator mode and settings
- Automatic file creation
-
Structured Logging: Zerolog-based logging (7fe863a)
- JSON and pretty console output
- Contextual fields for debugging
- Integration across orchestrator, adapters, and commands
- Maintains fmt.Fprintf for TUI display
-
Enhanced CLI: New commands and flags
agentpipe export- Export conversationsagentpipe resume- Resume saved conversationsagentpipe init- Interactive config wizard- --save-state, --state-file, --watch-config flags
π Code Quality Improvements
- Godoc comments on all exported types and functions
- 0 linting issues with golangci-lint
- Structured error types replacing fmt.Errorf
- Thread-safe implementations throughout
- Proper resource cleanup and leak prevention
v0.0.15 (October 14th, 2025)
New Agent Support
- GitHub Copilot CLI Integration: Full support for GitHub's Copilot terminal agent (
copilot)- Non-interactive mode support using
--promptflag - Automatic tool permission handling with
--allow-all-tools - Multi-model support (Claude Sonnet 4.5, GPT-5, etc.)
- Authentication detection and helpful error messages
- Subscription requirement validation
- Non-interactive mode support using
UX Improvements
- Graceful CTRL-C Handling: Interrupting a conversation now displays a session summary
- Total messages (agent + system)
- Total tokens used
- Total time spent (intelligently formatted: ms/s/m:s)
- Total conversation cost
- All messages are properly logged before exit
- Total Time Tracking in TUI: Statistics panel now shows cumulative time for all agent requests
Bug Fixes & Performance
- Resource Leak Fixes:
- Fixed timer leak in cursor adapter (using
time.NewTimerwith proper cleanup) - Message channel now properly closed on TUI exit
- Dropped messages now logged to stderr with counts
- Orchestrator goroutine lifecycle properly tracked with graceful shutdown
- Fixed timer leak in cursor adapter (using
v0.0.9 Updates
Agent Support
- Cursor CLI Integration: Full support for Cursor's AI agent (
cursor-agent)- Automatic authentication detection
- Intelligent retry logic for improved reliability
- Optimized timeout handling for cursor-agent's longer response times
- JSON stream parsing for real-time response streaming
- Robust error recovery and process management
v0.0.8 Features
TUI Improvements
- Real-time Activity Indicators: Visual feedback showing which agent is currently responding
- Enhanced Metrics Display:
- Response time shown in seconds with 1 decimal precision (e.g., 2.5s)
- Token count for each response
- Cost estimate per response (e.g., $0.0012)
- Total conversation cost tracking in Statistics panel
- Improved Message Formatting:
- Consolidated headers (timestamp and name only shown when speaker changes)
- Proper multi-paragraph message handling
- Clean spacing between messages
- No extra newlines between paragraphs from same speaker
- Configuration Improvements:
- TUI now properly honors all configuration settings
- Config file path displayed in Configuration panel
- Dual output support (logs to file while displaying in TUI)
- Metrics display controlled by
show_metricsconfig option
Agent & Orchestration
- Better Error Handling: Clearer error messages for agent failures and timeouts
- Improved Health Checks: More robust agent verification before starting conversations
- Cost Tracking: Automatic calculation and accumulation of API costs
- Metrics Pipeline: End-to-end metrics flow from orchestrator to TUI display
Key Improvements in Latest Version
Performance & Reliability
- Optimized Message Handling: Reduced memory usage and improved message rendering performance
- Better Concurrency: Proper goroutine management and channel handling
- Graceful Shutdowns: Clean termination of agents and proper resource cleanup
User Experience
- Intuitive Panel Navigation: Tab-based navigation between panels
- Real-time Feedback: Instant visual indicators for agent activity
- Clean Message Display: Smart consolidation of headers and proper paragraph formatting
- Cost Transparency: See exactly how much each conversation costs
Installation
Using Homebrew (macOS/Linux)
brew tap kevinelliott/tap
brew install agentpipe
Using the install script
curl -sSL https://raw.githubusercontent.com/kevinelliott/agentpipe/main/install.sh | bash
Using Go
go install github.com/kevinelliott/agentpipe@latest
Build from source
git clone https://github.com/kevinelliott/agentpipe.git
cd agentpipe
go build -o agentpipe .
Prerequisites
AgentPipe requires at least one AI CLI tool to be installed:
- Amp CLI -
amp- Install: See installation guide
- Authenticate: Follow Amp documentation
- Features: Autonomous coding, IDE integration, complex task execution
- Claude CLI -
claude - GitHub Copilot CLI -
copilot- Install:
npm install -g @github/copilot - Authenticate: Launch
copilotand use/logincommand - Requires: Node.js v22+, npm v10+, and active GitHub Copilot subscription
- Install:
- Cursor CLI -
cursor-agent- Install:
curl https://cursor.com/install -fsS | bash - Authenticate:
cursor-agent login
- Install:
- Gemini CLI -
gemini - Qwen CLI -
qwen - Codex CLI -
codex - Ollama -
ollama
Check which agents are available on your system:
agentpipe doctor
Quick Start
Simple conversation with command-line flags
# Start a conversation between Claude and Gemini
agentpipe run -a claude:Alice -a gemini:Bob -p "Let's discuss AI ethics"
# Use TUI mode with metrics for a rich experience
agentpipe run -a claude:Poet -a gemini:Scientist --tui --metrics
# Configure conversation parameters
agentpipe run -a claude:Agent1 -a gemini:Agent2 \
--mode reactive \
--max-turns 10 \
--timeout 45 \
--prompt "What is consciousness?"
Using configuration files
# Run with a configuration file
agentpipe run -c examples/simple-conversation.yaml
# Run a debate between three agents
agentpipe run -c examples/debate.yaml --tui
# Brainstorming session with multiple agents
agentpipe run -c examples/brainstorm.yaml
Configuration
YAML Configuration Format
version: "1.0"
agents:
- id: agent-1
type: claude # Agent type (claude, gemini, qwen, etc.)
name: "Friendly Assistant"
prompt: "You are a helpful and friendly assistant."
announcement: "Hello everyone! I'm here to help!"
model: claude-3-sonnet # Optional: specific model
temperature: 0.7 # Optional: response randomness
max_tokens: 1000 # Optional: response length limit
- id: agent-2
type: gemini
name: "Technical Expert"
prompt: "You are a technical expert who loves explaining complex topics."
announcement: "Technical Expert has joined the chat!"
temperature: 0.5
orchestrator:
mode: round-robin # Conversation mode
max_turns: 10 # Maximum conversation turns
turn_timeout: 30s # Timeout per agent response
response_delay: 2s # Delay between responses
initial_prompt: "Let's start our discussion!"
logging:
enabled: true # Enable chat logging
chat_log_dir: ~/.agentpipe/chats # Custom log path (optional)
show_metrics: true # Display response metrics in TUI (time, tokens, cost)
log_format: text # Log format (text or json)
Conversation Modes
- round-robin: Agents speak in a fixed rotation
- reactive: Agents respond based on who spoke last
- free-form: Agents decide when to participate
Commands
agentpipe run
Start a conversation between agents.
Flags:
-c, --config: Path to YAML configuration file-a, --agents: List of agents (format:type:name)-m, --mode: Conversation mode (default: round-robin)--max-turns: Maximum conversation turns (default: 10)--timeout: Response timeout in seconds (default: 30)--delay: Delay between responses in seconds (default: 1)-p, --prompt: Initial conversation prompt-t, --tui: Use enhanced TUI interface with panels and user input--log-dir: Custom path for chat logs (default: ~/.agentpipe/chats)--no-log: Disable chat logging--metrics: Display response metrics (duration, tokens, cost) in TUI--skip-health-check: Skip agent health checks (not recommended)--health-check-timeout: Health check timeout in seconds (default: 5)--save-state: Save conversation state to file on completion--state-file: Custom state file path (default: auto-generated)--watch-config: Watch config file for changes and reload (development mode)
agentpipe doctor
Check which AI CLI tools are installed and available.
agentpipe doctor
agentpipe export
Export conversation from a state file to different formats.
# Export to JSON
agentpipe export state.json --format json --output conversation.json
# Export to Markdown
agentpipe export state.json --format markdown --output conversation.md
# Export to HTML (includes styling)
agentpipe export state.json --format html --output conversation.html
Flags:
--format: Export format (json, markdown, html)--output: Output file path
agentpipe resume
Resume a saved conversation from a state file.
# List all saved conversations
agentpipe resume --list
# View a saved conversation
agentpipe resume ~/.agentpipe/states/conversation-20231215-143022.json
# Resume and continue (future feature)
agentpipe resume state.json --continue
Flags:
--list: List all saved conversation states--continue: Continue the conversation (planned feature)
agentpipe init
Interactive wizard to create a new AgentPipe configuration file.
agentpipe init
Creates a YAML config file with guided prompts for:
- Conversation mode selection
- Agent configuration
- Orchestrator settings
- Logging preferences
Examples
Cursor and Claude Collaboration
# Save as cursor-claude-team.yaml
version: "1.0"
agents:
- id: cursor-dev
type: cursor
name: "Cursor Developer"
prompt: "You are a senior developer who writes clean, efficient code."
- id: claude-reviewer
type: claude
name: "Claude Reviewer"
prompt: "You are a code reviewer who ensures best practices and identifies potential issues."
orchestrator:
mode: round-robin
max_turns: 6
initial_prompt: "Let's design a simple REST API for a todo list application."
Poetry vs Science Debate
# Save as poetry-science.yaml
version: "1.0"
agents:
- id: poet
type: claude
name: "The Poet"
prompt: "You speak in beautiful metaphors and see the world through an artistic lens."
temperature: 0.9
- id: scientist
type: gemini
name: "The Scientist"
prompt: "You explain everything through logic, data, and scientific principles."
temperature: 0.3
orchestrator:
mode: round-robin
initial_prompt: "Is love just chemistry or something more?"
Run with: agentpipe run -c poetry-science.yaml --tui
Creative Brainstorming with Metrics
agentpipe run \
-a claude:IdeaGenerator \
-a gemini:CriticalThinker \
-a qwen:Implementer \
--mode free-form \
--max-turns 15 \
--metrics \
--tui \
-p "How can we make education more engaging?"
When metrics are enabled, you'll see:
- Response time for each agent (e.g., "2.3s")
- Token usage per response (e.g., "150 tokens")
- Cost estimate per response (e.g., "$0.0023")
- Total conversation cost in the Statistics panel
TUI Interface
The enhanced TUI provides a rich, interactive experience for managing multi-agent conversations:
Layout
The TUI is divided into multiple panels:
- Agents Panel (Left): Shows all connected agents with real-time status indicators
- Chat Panel (Center): Displays the conversation with color-coded messages
- Topic Panel (Top Right): Shows the initial conversation prompt
- Statistics Panel (Right): Displays turn count, agent statistics, and total conversation cost
- Configuration Panel (Right): Shows active settings and config file path
- User Input Panel (Bottom): Allows you to participate in the conversation
Visual Features
- Agent Status Indicators: Green dot (π’) for active/responding, grey dot (β«) for idle
- Color-Coded Messages: Each agent gets a unique color for easy tracking
- Consolidated Headers: Message headers only appear when the speaker changes
- Metrics Display: Response time (seconds), token count, and cost shown inline when enabled
- Multi-Paragraph Support: Properly formatted multi-line agent responses
Controls
Tab: Switch between panels (Agents, Chat, User Input)ββ: Navigate in active panelEnter: Send message when in User Input paneli: Show agent info modal (when in Agents panel)Ctrl+Corq: QuitPageUp/PageDown: Scroll conversation- Active agent indicators: π’ (responding) / β« (idle)
Development
Building from Source
# Clone the repository
git clone https://github.com/kevinelliott/agentpipe.git
cd agentpipe
# Build the binary
go build -o agentpipe .
# Or build with version information
VERSION=v0.0.7 make build
# Run tests
go test ./...
Project Structure
agentpipe/
βββ cmd/ # CLI commands
β βββ root.go # Root command
β βββ run.go # Run conversation command
β βββ doctor.go # Doctor diagnostic command
β βββ export.go # Export conversations
β βββ resume.go # Resume conversations
β βββ init.go # Interactive configuration wizard
βββ pkg/
β βββ agent/ # Agent interface and registry
β βββ adapters/ # Agent implementations (7 adapters)
β βββ config/ # Configuration handling
β β βββ watcher.go # Config hot-reload support
β βββ conversation/ # Conversation state management
β β βββ state.go # Save/load conversation states
β βββ errors/ # Structured error types
β βββ export/ # Export to JSON/Markdown/HTML
β βββ log/ # Structured logging (zerolog)
β βββ logger/ # Chat logging and output
β βββ metrics/ # Prometheus metrics
β β βββ metrics.go # Metrics collection
β β βββ server.go # HTTP metrics server
β βββ middleware/ # Message processing pipeline
β β βββ middleware.go # Core middleware pattern
β β βββ builtin.go # Built-in middleware
β βββ orchestrator/ # Conversation orchestration
β βββ ratelimit/ # Token bucket rate limiting
β βββ tui/ # Terminal UI
β βββ utils/ # Utilities (tokens, costs)
βββ docs/ # Documentation
β βββ architecture.md
β βββ contributing.md
β βββ development.md
β βββ troubleshooting.md
β βββ docker.md
βββ examples/ # Example configurations
β βββ simple-conversation.yaml
β βββ brainstorm.yaml
β βββ middleware.yaml
β βββ prometheus-metrics.yaml
βββ test/
β βββ integration/ # End-to-end tests
β βββ benchmark/ # Performance benchmarks
βββ Dockerfile # Multi-stage production build
βββ docker-compose.yml # Docker Compose configuration
βββ main.go
Adding New Agent Types
- Create a new adapter in
pkg/adapters/ - Implement the
Agentinterface - Register the factory in
init()
type MyAgent struct {
agent.BaseAgent
}
func init() {
agent.RegisterFactory("myagent", NewMyAgent)
}
Advanced Features
Prometheus Metrics & Monitoring
AgentPipe includes comprehensive Prometheus metrics for production monitoring:
// Enable metrics in your code
import "github.com/kevinelliott/agentpipe/pkg/metrics"
// Start metrics server
server := metrics.NewServer(metrics.ServerConfig{Addr: ":9090"})
go server.Start()
// Set metrics on orchestrator
orch.SetMetrics(metrics.DefaultMetrics)
Available Metrics:
agentpipe_agent_requests_total- Request counter by agent and statusagentpipe_agent_request_duration_seconds- Request duration histogramagentpipe_agent_tokens_total- Token usage by type (input/output)agentpipe_agent_cost_usd_total- Estimated costs in USDagentpipe_agent_errors_total- Error counter by typeagentpipe_active_conversations- Current active conversationsagentpipe_conversation_turns_total- Total turns by modeagentpipe_message_size_bytes- Message size distributionagentpipe_retry_attempts_total- Retry counteragentpipe_rate_limit_hits_total- Rate limit hits
Endpoints:
http://localhost:9090/metrics- Prometheus metrics (OpenMetrics format)http://localhost:9090/health- Health checkhttp://localhost:9090/- Web UI with documentation
See examples/prometheus-metrics.yaml for complete configuration, Prometheus queries, Grafana dashboard setup, and alerting rules.
Docker Support
Run AgentPipe in Docker for production deployments:
# Build image
docker build -t agentpipe:latest .
# Run with docker-compose (includes metrics server)
docker-compose up
# Run standalone
docker run -v ~/.agentpipe:/root/.agentpipe agentpipe:latest run -c /config/config.yaml
Features:
- Multi-stage build (~50MB final image)
- Health checks included
- Volume mounts for configs and logs
- Prometheus metrics exposed on port 9090
- Production-ready with non-root user
See docs/docker.md for complete Docker documentation.
Middleware Pipeline
Extend AgentPipe with custom message processing:
// Add built-in middleware
orch.AddMiddleware(middleware.LoggingMiddleware())
orch.AddMiddleware(middleware.MetricsMiddleware())
orch.AddMiddleware(middleware.ContentFilterMiddleware(config))
// Or use defaults
orch.SetupDefaultMiddleware()
// Create custom middleware
custom := middleware.NewTransformMiddleware("uppercase",
func(ctx *MessageContext, msg *Message) (*Message, error) {
msg.Content = strings.ToUpper(msg.Content)
return msg, nil
})
orch.AddMiddleware(custom)
Built-in Middleware:
LoggingMiddleware- Structured loggingMetricsMiddleware- Performance trackingContentFilterMiddleware- Content validation and filteringSanitizationMiddleware- Message sanitizationEmptyContentValidationMiddleware- Empty message rejectionRoleValidationMiddleware- Role validationErrorRecoveryMiddleware- Panic recovery
See examples/middleware.yaml for complete examples.
Rate Limiting
Configure rate limits per agent:
agents:
- id: claude
type: claude
rate_limit: 10 # 10 requests per second
rate_limit_burst: 5 # Burst capacity of 5
Uses token bucket algorithm with:
- Configurable rate and burst capacity
- Thread-safe implementation
- Automatic rate limit hit tracking in metrics
Conversation State Management
Save and resume conversations:
# Save conversation state on completion
agentpipe run -c config.yaml --save-state
# List saved conversations
agentpipe resume --list
# View saved conversation
agentpipe resume ~/.agentpipe/states/conversation-20231215-143022.json
# Export to different formats
agentpipe export state.json --format html --output report.html
State files include:
- Full conversation history
- Configuration used
- Metadata (turns, duration, timestamps)
- Agent information
Config Hot-Reload (Development Mode)
Enable config file watching for rapid development:
agentpipe run -c config.yaml --watch-config
Changes to the config file are automatically detected and reloaded without restarting the conversation.
Troubleshooting
Agent Health Check Failed
If you encounter health check failures:
- Verify the CLI is properly installed:
which <agent-name> - Check if the CLI requires authentication or API keys
- Try running the CLI manually to ensure it works
- Use
--skip-health-checkflag as a last resort (not recommended)
GitHub Copilot CLI Issues
The GitHub Copilot CLI has specific requirements:
- Authentication: Run
copilotin interactive mode and use/logincommand - Subscription Required: Requires an active GitHub Copilot subscription
- Model Selection: Default is Claude Sonnet 4.5; use
modelconfig option to specify others - Node.js Requirements: Requires Node.js v22+ and npm v10+
- Check Status: Run
copilot --helpto verify installation
Cursor CLI Specific Issues
The Cursor CLI (cursor-agent) has some unique characteristics:
- Authentication Required: Run
cursor-agent loginbefore first use - Longer Response Times: Cursor typically takes 10-20 seconds to respond (AgentPipe handles this automatically)
- Process Management: cursor-agent doesn't exit naturally; AgentPipe manages process termination
- Check Status: Run
cursor-agent statusto verify authentication - Timeout Errors: If you see timeout errors, ensure you're authenticated and have a stable internet connection
Qwen Code CLI Issues
The Qwen Code CLI uses a different interface than other agents:
- Use
qwen --prompt "your prompt"for non-interactive mode - The CLI may open an interactive session if not properly configured
- Full documentation: https://github.com/QwenLM/qwen-code
Gemini Model Not Found
If you get a 404 error with Gemini:
- Check your model name in the configuration
- Ensure you have access to the specified model
- Try without specifying a model to use the default
Chat Logs Location
Chat logs are saved by default to:
- macOS/Linux:
~/.agentpipe/chats/ - Windows:
%USERPROFILE%\.agentpipe\chats\
You can override this with --log-path or disable logging with --no-log.
License
MIT License
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
Documentation
ΒΆ
There is no documentation for this package.
Directories
ΒΆ
| Path | Synopsis |
|---|---|
|
Package main demonstrates the export functionality
|
Package main demonstrates the export functionality |
|
internal
|
|
|
pkg
|
|
|
agent
Package agent provides the core interfaces and types for AI agent communication.
|
Package agent provides the core interfaces and types for AI agent communication. |
|
config
Package config provides configuration management for AgentPipe.
|
Package config provides configuration management for AgentPipe. |
|
conversation
Package conversation provides conversation state management for AgentPipe.
|
Package conversation provides conversation state management for AgentPipe. |
|
export
Package export provides functionality to export conversations to different formats.
|
Package export provides functionality to export conversations to different formats. |
|
log
Package log provides structured logging using zerolog.
|
Package log provides structured logging using zerolog. |
|
metrics
Package metrics provides Prometheus metrics for AgentPipe.
|
Package metrics provides Prometheus metrics for AgentPipe. |
|
middleware
Package middleware provides a flexible middleware pattern for message processing.
|
Package middleware provides a flexible middleware pattern for message processing. |
|
orchestrator
Package orchestrator manages multi-agent conversations with different orchestration modes.
|
Package orchestrator manages multi-agent conversations with different orchestration modes. |
|
ratelimit
Package ratelimit provides token bucket rate limiting for agent requests.
|
Package ratelimit provides token bucket rate limiting for agent requests. |