Goose — User Demand Report
Week: 2026-W14 Generated: 2026-04-03 Issues analyzed: 32 (28 included) Need clusters: 6
Top 10 User Needs
| Rank | Need | Issues | Score | Category | Examples |
|---|---|---|---|---|---|
| 1 | Improve CLI stability, usability, and local model management | 10 | 2.2 | Developer Experience | #8282, #8275, #8274 |
| 2 | Improving AI Agent Reliability, Configuration, and Extensibility | 9 | 1.9 | Developer Experience | #8270, #8265, #8264 |
| 3 | Comprehensive Documentation Updates, Fixes, and Quickstart Enhancements | 3 | 1.0 | Documentation | #8269, #8226, #8195 |
| 4 | Improve Cloud AI API Version Routing Compatibility | 2 | 0.0 | Integration | #8277, #8236 |
| 5 | Resource Monitoring and Process Lifecycle Management | 2 | 0.0 | Performance | #8266, #8229 |
| 6 | Flexible LLM Provider Settings and Per-Agent Assignments | 2 | 0.0 | Configuration | #8203, #8187 |
Rising Needs
| Need | Rising Score | This Week | Category |
|---|---|---|---|
| Improve CLI stability, usability, and local model management | 11.0x | 10 | Developer Experience |
| Improving AI Agent Reliability, Configuration, and Extensibility | 10.0x | 9 | Developer Experience |
| Comprehensive Documentation Updates, Fixes, and Quickstart Enhancements | 4.0x | 3 | Documentation |
| Improve Cloud AI API Version Routing Compatibility | 3.0x | 2 | Integration |
| Resource Monitoring and Process Lifecycle Management | 3.0x | 2 | Performance |
Category Breakdown
- Developer Experience: 2 clusters
- Documentation: 1 clusters
- Integration: 1 clusters
- Performance: 1 clusters
- Configuration: 1 clusters
All Need Clusters
1. Improve CLI stability, usability, and local model management
Users want enhancements to the command-line interface, including better model management, crash fixes, log cleanup, and improved interactive session features. These changes aim to create a more reliable, user-friendly terminal experience that minimizes friction and clarifies output during local AI workflows.
- Volume: 10 issues (8 open, 2 closed)
- Demand Score: 2.2
- Avg Reactions: 0.1 | Avg Comments: 0.4
- Example issues: #8282, #8275, #8274, #8273, #8272
2. Improving AI Agent Reliability, Configuration, and Extensibility
Users need critical bug fixes for runtime crashes, authentication issues, and model regressions to maintain reliable AI agent performance. They also require enhanced configuration controls, automated maintenance workflows, and custom security hooks to better adapt the extension to complex development environments.
- Volume: 9 issues (8 open, 1 closed)
- Demand Score: 1.9
- Avg Reactions: 0 | Avg Comments: 0.2
- Example issues: #8270, #8265, #8264, #8262, #8248
3. Comprehensive Documentation Updates, Fixes, and Quickstart Enhancements
Users are requesting comprehensive updates to the project documentation to cover new features, correct outdated references, and fix copy-paste errors. Accurate and current documentation is crucial to prevent configuration mistakes and streamline the onboarding process for new developers.
- Volume: 3 issues (3 open, 0 closed)
- Demand Score: 1.0
- Avg Reactions: 0 | Avg Comments: 0.7
- Example issues: #8269, #8226, #8195
4. Improve Cloud AI API Version Routing Compatibility
Users need the system to correctly route API requests and handle version query parameters for external cloud AI providers like GCP and Azure. Implementing conditional parameter exclusion and fixing global endpoint routing will resolve compatibility issues across different API versions. This ensures reliable integration with multiple cloud-hosted AI services.
- Volume: 2 issues (2 open, 0 closed)
- Demand Score: 0.0
- Avg Reactions: 0 | Avg Comments: 0
- Example issues: #8277, #8236
5. Resource Monitoring and Process Lifecycle Management
Users want integrated tools to track LLM token usage and costs while ensuring system resources are properly cleaned up. They are requesting automatic process termination at the end of sessions to eliminate memory leaks and prevent resource exhaustion. These improvements aim to enhance application stability and provide clear visibility into operational expenses.
- Volume: 2 issues (1 open, 1 closed)
- Demand Score: 0.0
- Avg Reactions: 0 | Avg Comments: 0
- Example issues: #8266, #8229
6. Flexible LLM Provider Settings and Per-Agent Assignments
Users want the ability to assign specific LLM models and API keys to individual agents within a single provider setup. They also need the flexibility to override base URLs for built-in providers to support custom deployments or proxy routing. These enhancements would improve environment isolation, cost tracking, and infrastructure adaptability.
- Volume: 2 issues (1 open, 1 closed)
- Demand Score: 0.0
- Avg Reactions: 0 | Avg Comments: 0
- Example issues: #8203, #8187
This report analyzes public GitHub issues only. It represents a signal from public issue discussions, not the full user base.
Generated by ReadYourUsers