Fallback benchmark comparison to dry-run when secrets are unavailable#18
Merged
sharpninja merged 3 commits intomainfrom Mar 21, 2026
Merged
Conversation
Co-authored-by: sharpninja <16146732+sharpninja@users.noreply.github.com> Agent-Logs-Url: https://github.com/sharpninja/graphrag/sessions/cd6db3fa-53c3-4678-a5ee-0bc3ee2bdb6f
Co-authored-by: sharpninja <16146732+sharpninja@users.noreply.github.com> Agent-Logs-Url: https://github.com/sharpninja/graphrag/sessions/cd6db3fa-53c3-4678-a5ee-0bc3ee2bdb6f
Copilot
AI
changed the title
[WIP] Fix build failure on benchmark pipeline
Fallback benchmark comparison to dry-run when secrets are unavailable
Mar 21, 2026
sharpninja
approved these changes
Mar 21, 2026
|
You have reached your Codex usage limits for code reviews. You can see your limits in the Codex usage dashboard. |
There was a problem hiding this comment.
Pull request overview
Updates the Benchmark Comparison GitHub Actions workflow to automatically fall back to a --dry-run benchmark execution path when required benchmark secrets are not available, ensuring JSON outputs and the comparison report are still produced.
Changes:
- Adds a “Select benchmark mode” step that detects whether required secret-backed environment variables are present and sets
BENCHMARK_EXTRA_ARGaccordingly. - Passes
${BENCHMARK_EXTRA_ARG}into both Python and .NET benchmark invocations so they can run in full mode or--dry-run. - Writes a clear mode indicator into the job step summary (full vs dry-run).
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Description
The benchmark comparison workflow was failing before it could produce artifacts when required benchmark secrets were not configured. This change makes the pipeline degrade to a dry-run path in that case, while preserving full benchmark execution when the environment is fully configured.
Related Issues
Addresses the benchmark pipeline build failure.
Proposed Changes
Workflow mode selection
Select benchmark modestep to.github/workflows/benchmark-comparison.yml.Graceful fallback
scripts/benchmark_smoke.pywith--dry-runwhen benchmark secrets are missing.Artifact/report continuity
Checklist
Additional Notes
The failure mode was not in report generation; the Python benchmark step was exiting during fixture execution because secret-backed model/search configuration was unresolved. This change keeps the workflow useful in forks and CI environments where those secrets are intentionally absent.
📍 Connect Copilot coding agent with Jira, Azure Boards or Linear to delegate work to Copilot in one click without leaving your project management tool.