[prompt-analysis] Copilot PR Prompt Analysis - December 3, 2025 #5394
Closed
Replies: 1 comment
-
|
⚓ Avast! This discussion be marked as outdated by Copilot PR Prompt Pattern Analysis. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Summary
Analysis of Copilot-generated PRs over the last 30 days reveals a 79.8% merge success rate across 997 completed PRs. The data shows that prompts with code references (98% in merged vs 87% in closed), file-specific mentions (86% vs 78%), and test-related keywords achieve the highest success rates.
Analysis Period: Last 30 days (Nov 3 - Dec 3, 2025)
Total PRs: 1,000 | Merged: 796 (79.8%) | Closed: 201 (20.2%) | Open: 3 (0.3%)
Full Analysis Report
Prompt Categories and Success Rates
All categories show similar success rates (79.7-80.9%), suggesting category choice is less important than prompt quality.
Prompt Characteristics Analysis
✅ Successful Prompt Patterns (Merged PRs)
Common characteristics:
Key differentiators for merged PRs:
❌ Unsuccessful Prompt Patterns (Closed PRs)
Common characteristics:
Warning signs observed:
Example Comparisons
✅ Successful Prompt Example
PR #5373 - Document safe-outputs requirements in schema with $comment
Why it succeeded:
safe-outputs)✅ Successful Prompt Example
PR #5372 - Convert stale repository identifier to agentic workflow
Why it succeeded:
❌ Unsuccessful Prompt Example
PR #5377 - Update workflow to review only public repos [WIP]
Why it failed:
❌ Unsuccessful Prompt Example
PR #5355 - Address network firewall warnings [WIP]
Why it failed:
Key Insights
Based on 997 completed PRs, the data reveals three critical success factors:
1. Code Specificity Matters Most (+11% success difference)
Merged PRs are 11% more likely to include code references (98% vs 87%). Using backticks for code, function names, and technical terms signals concrete implementation focus.
Pattern: Successful prompts show the code, not just describe it.
2. File-Level Detail Improves Outcomes (+8% success difference)
Merged PRs mention specific files 8% more often (86% vs 78%). Referencing
.go,.js,.yaml, or exact file paths demonstrates understanding of the codebase structure.Pattern: "Fix the authentication handler" → "Fix
pkg/auth/handler.goauthentication logic"3. Error Context Provides Direction (+6% success difference)
While only 50% of merged PRs include error messages, they still outperform closed PRs (44%). Including error text, stack traces, or bug descriptions helps Copilot understand the problem domain.
Pattern: Including the actual error message guides implementation better than generic "fix bug" requests.
Recommendations
Based on this analysis, follow these best practices for Copilot PR prompts:
✅ DO: Write Specific, Code-Focused Prompts
Include code references: Use backticks for functions, variables, error messages
validateInput()function inpkg/parser/validator.go"Reference specific files: Mention exact file paths when known
pkg/parser/schemas/workflow.json"Provide error context: Include actual error messages or bug descriptions
Detail the implementation: Describe what should change, not just what's broken
❌ AVOID: Generic or Incomplete Prompts
📝 Prompt Template for Success
Statistical Summary
Prompt Quality Indicators (Merged vs Closed):
Key Takeaway: Specificity (code/file references) matters more than length or category.
Category Performance Insights
While all categories show similar success rates (79.7-80.9%), slight variations exist:
Insight: The type of work matters less than how you describe it. A well-written feature prompt outperforms a vague bug fix prompt.
Methodology Notes
Data Collection:
Categorization:
Limitations:
Analysis Date: December 3, 2025
Generated by: Copilot PR Prompt Analysis Workflow
Run ID: §19888397028
Beta Was this translation helpful? Give feedback.
All reactions