We are proud to announce the SoccerNet Challenge 2026 - Visual Question Answering (VQA), building upon our previous work presented in ⚽️"Multi-Agent System for Comprehensive Soccer Understanding"⚽️ (MM'25). This challenge aims to push the boundaries of multimodal reasoning in multi-modal soccer understanding through a comprehensive question-answering framework. Here is the challenge page, more details about the SoccerAgent project could be found here.
More details could be found at:
- 📑 Paper Relevant Links: Paper ⋅ WebPage ⋅ Benchmark ⋅ Database
- 🏆 2026 SoccerNet Challenge - VQA: Benchmark (Test & Challenge) ⋅ Eval (Test) ⋅ Eval (Challenge)
No.1 in Challenge Phase: $1,000💰 (Sponsored by KNQ Technology)
Participants will tackle close-ended questions from 14 distinct (13 from SoccerAgent paper, 1 newly added) soccer understanding tasks categorized into three main modalities:
- Background Knowledge QA: Questions about players, teams, referees, and venues
- Match Situation QA: Queries about specific match events and statistics
- Camera Status Classification: Identifying camera positions in broadcast images
- Jersey Number Relevant QA: Detecting and recognizing player jersey numbers
- Score/Time Relevant QA: Extracting game time and scoreboard information
- Background Knowledge Image QA: Visual questions about soccer entities
- Game State QA (Newly added): Questions counting the offensing/defencing players
- Camera Status Switching: Classifying camera transition types in video clips
- Replay Grounding: Identifying which action is being replayed
- Action Classification: Recognizing soccer-specific actions and events
- Commentary Generation: Producing textual descriptions from video content
- Commentary Relevant QA: Questions combining player background and match commentary
- Jersey Color Relevant QA: Questions incorporating jersey color information
- Multi-view Foul Recognition: Analyzing fouls from multiple camera angles
With submission and evaluation all on CodaBench pages: Test & Challenge
As for this close-ended QA task, we directly use the accuracy as the evaluation metric:
- Challenge Set: 500 held-out multi-modal soccer questions (500 questions in testset are not counted for ranking)
- Evaluation Server: Automated scoring system with leaderboard
- Submission Format: Zip file containing a JSON file with predicted answer
From the challenge dataset page here, you can find the train/valid/test/challenge set of this challenge. With each .zip file contains a .json file and a materials folder, where .json file contains the questions (challenge.zip has no answers) and materials folder contains the required picture/videos in the questions.
Such train.zip and valid.zip is split from the benchmark SoccerBench, which is the largest multimodal soccer understanding QA benchmark introduced by our previous work of SoccerAgent. In the train/valid dataset, you could find the soccer understanding QA pairs featuring:
- Around 10k QA pairs across 14 tasks
- Balanced distribution across modalities and difficulty levels
- Multi-choice format with one correct answer and three distractors
- Professional annotations from soccer experts
Also, there are many of the reference sources (with links) are recommended to solve the questions in this challenge:
- SoccerWiki: Large-scale knowledge base with 9,471 players, 266 teams, 202 referees
- MatchTime & SoccerReplay-1988: Commentary and event annotation datasets
- SoccerNet Series Dataset: There are many different soccer understanding tasks from Soccernet community (e.g. Action spotting, foul recognition, and jersey number datasets .etc).
We provide baseline pipelines for such comprehensive multi-modal soccer understanding task:
The first one is to use VLMs: We provide baseline implementations for two models: Qwen2.5VL and GPT4o. You can run VQA tests with either model by passing the appropriate arguments to the script ./baseline.py.
Here is how you can run the script:
python ./baseline.py \
--model <model_type> \
--input_file <input_json_file> \
--output_file <output_json_file> \
--materials_folder <materials_directory> \
[--model_path <qwen_model_path>] \
[--api_key <gpt_api_key>]--model: Specify the model to use. Options areqwenfor Qwen2.5VL andgptfor GPT4o.--input_file: Path to the input JSON file with a series of questions and options.--output_file: Path to the output JSON file where results (id and answer) will be saved.--materials_folder: Path to the folder containing materials used in the QA task.--api_key: (Required if using GPT4o) Your API key for accessing GPT4o services.--model_path: (Required if using Qwen2.5VL) Path to the local Qwen2.5VL model.
After running the script, you'll get an output JSON file at the specified location. Rename it to metadata.json, compress it as a ZIP file, and upload it to the Codabench challenge to obtain the baseline accuracy.
Also, you can reference the SoccerAgent Pipeline:, the implementation is shown here. You can modify the pipeline and toolbox to fit your settings. Or you can just utilize the EXECUTE_TOOL_CHAIN function in ./multiagent_platform.py to solve the questions one by one. To be noted, SoccerAgent requires lots of installations and distributions of different work, feel free to contact and you can modify them as you want.
- Open to individuals and teams worldwide
- Organizers and their immediate collaborators are ineligible
- Commercial and academic teams are equally welcome
- The right of interpretation belongs to the organizer
- Computational Limits: No restrictions on model size or compute
- Framework Freedom: Any ML framework or toolkit allowed
- Prediction Files: JSON format with specified structure
- Technical Report: If you are the winner of the prize, 4-page paper (CVPR format) describing methodology
- Demo Video: If you are the winner of the prize, a short presentation of the approach is required
Details could be find in codabench page before, you need to zip a json file with the according format.
submission.zip/
└─── metadata.json
The json file format is shown as the following for all the questions:
[
{
"id": 1,
"Answer": "O3"
},
{
"id": 2,
"Answer": "O4"
},
{
"id": 3,
"Answer": "O1"
},
...
]
An example of such JSON file could be found here
| Date | Milestone | Description |
|---|---|---|
| Oct 15, 2025 | Challenge Launch | Release of test/challenge data |
| OCt 15, 2025 | Evaluation Server Open | Submission system activated |
| Apr 24, 2026 | Final Submission Deadline | Last chance for leaderboard updates |
| Jun TBD, 2026 | Presentation | Winner give presentation @CVSports |
This challenge aims to:
- Advance the state-of-the-art in soccer video understanding
- Establish standardized evaluation protocols for sports AI
- Foster collaboration between computer vision and sports analytics communities
- Create publicly available benchmarks for future research
- Bridge the gap between academic research and real-world applications
Organized in collaboration with the CVSports Workshop at CVPR 2026. Based on the framework established in "Multi-Agent System for Comprehensive Soccer Understanding" (MM'25).
For questions and support, please visit 2026 SoccerNet Challenge Website or join our Discord Group.

