[prompt-clustering] Copilot Agent Prompt Clustering Analysis - 2025-12-01 #5232
Closed
Replies: 1 comment
-
|
⚓ Avast! This discussion be marked as outdated by Copilot Agent Prompt Clustering Analysis. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
🔬 Copilot Agent Prompt Clustering Analysis
Daily NLP-based clustering analysis of copilot agent task prompts
Analysis Date: 2025-12-01
Executive Summary
Analyzed 1,373 copilot agent tasks using NLP clustering techniques (TF-IDF + K-means).
The analysis identified 8 distinct task categories with an overall success rate of 75.3% (merged PRs).
Key Findings:
Full Clustering Analysis Report
Overall Statistics
Methodology
NLP Pipeline:
Cluster Analysis
Cluster 1: Documentation
Size: 408 tasks (29.7% of total)
Success Metrics:
Code Change Profile:
Top Keywords: github, workflows, documentation, cli, command, mcp, copilot, code
Characteristics: This cluster represents documentation tasks.
Example PRs: #2097, #2099, #2100
Cluster 2: Bug Fix
Size: 233 tasks (17.0% of total)
Success Metrics:
Code Change Profile:
Top Keywords: workflow, agentic, workflows, run, file, github, issue, agent
Characteristics: This cluster represents bug fix tasks.
High success rate (84.1%) indicates these tasks are well-suited for the agent.
Example PRs: #2103, #2104, #2107
Cluster 3: Update/Enhancement
Size: 202 tasks (14.7% of total)
Success Metrics:
Code Change Profile:
Top Keywords: output, safe, test, add, tests, update, run, changes
Characteristics: This cluster represents update/enhancement tasks.
Example PRs: #2111, #2127, #2137
Cluster 4: Documentation
Size: 140 tasks (10.2% of total)
Success Metrics:
Code Change Profile:
Top Keywords: copilot, servers, docs, mcp servers, aw make copilot, aw make, gh aw make, model context protocol
Characteristics: This cluster represents documentation tasks.
Example PRs: #2209, #2283, #2284
Cluster 5: General Development
Size: 135 tasks (9.8% of total)
Success Metrics:
Code Change Profile:
Top Keywords: copilot, thoughts, survey, thoughts copilot coding, share thoughts copilot, thoughts copilot, input share, love
Characteristics: This cluster represents general development tasks.
Example PRs: #2128, #2347, #2382
Cluster 6: Performance
Size: 129 tasks (9.4% of total)
Success Metrics:
Code Change Profile:
Top Keywords: set, coding, coding agent, copilot, agent, works faster, let copilot coding, higher quality
Characteristics: This cluster represents performance tasks.
High success rate (81.4%) indicates these tasks are well-suited for the agent.
Example PRs: #2254, #2282, #2285
Cluster 7: Documentation
Size: 78 tasks (5.7% of total)
Success Metrics:
Code Change Profile:
Top Keywords: files, workflow, github, command, documentation, file, workflows, agentic
Characteristics: This cluster represents documentation tasks.
High success rate (80.8%) indicates these tasks are well-suited for the agent.
Example PRs: #2108, #2109, #2151
Cluster 8: Feature Addition
Size: 48 tasks (3.5% of total)
Success Metrics:
Code Change Profile:
Top Keywords: make, work, changes, add, code, works faster, work set, use
Characteristics: This cluster represents feature addition tasks.
Lower success rate (8.3%) suggests these tasks may be more challenging.
Example PRs: #2101, #2110, #2126
Success Rate by Cluster
Sample Task Data
Representative sample of tasks from each cluster:
Key Findings
1. Documentation Tasks Dominate
626 tasks (45.6%) across 3 clusters involve documentation work.
This includes updating documentation, adding guides, improving CLI command docs, and MCP server documentation.
2. Bug Fixes Have High Success Rates
Bug fix tasks achieve 84.1% success rate,
suggesting the agent is particularly effective at targeted fixes with clear objectives.
3. Complex Feature Additions Are Challenging
The 'Feature Addition' cluster has the lowest success rate at 8.3%.
Tasks in this cluster involve: make, work, changes, add, code.
This suggests that open-ended feature additions may need more specific requirements or guidance.
Recommendations
Based on the clustering analysis, here are actionable recommendations:
1. Optimize for High-Success Task Types
Focus on task types with proven success rates:
2. Improve Prompts for Low-Success Clusters
Tasks in these categories need better-defined requirements:
3. Leverage Documentation Patterns
Since documentation tasks are most common and have solid success rates (76.7%),
create reusable prompt templates for:
4. Break Down Complex Features
Tasks involving complex features show lower success rates. Consider:
Methodology: TF-IDF vectorization with K-means clustering (k=8)
Data Coverage: 1,373 copilot agent PRs from githubnext/gh-aw
References:
Beta Was this translation helpful? Give feedback.
All reactions