iWISDM, short for instructed-Virtual VISual Decision Making, is a virtual environment capable of generating a limitless array of vision-language tasks with varying complexity.
It is a toolkit designed to evaluate the ability of multimodal models to follow instructions in visual tasks. It builds on the compositional nature of human behavior, and the fact that complex tasks are often constructed by combining smaller task units together in time.
iWISDM encompasses a broad spectrum of subtasks that engage executive functions such as inhibition of action, working memory, attentional set, task switching, and schema generalization.
It is also a scalable and extensible framework which allows users to easily define their own task space and stimuli dataset.
Using iWISDM, we have compiled four distinct benchmarks of increasing complexity levels for evaluation of large multi-modal models.
Below is an example of the generated tasks:
These datasets can be generated from /benchmarking or downloaded at: iWISDM_benchsets.tar.gz
iWISDM inherits several classes from COG (github.com/google/cog) to build task graphs. For convenience, we have also pre-implemented several commonly used cognitive tasks in task_bank.py.
https://arxiv.org/submit/5678755/view
To install the iWISDM package, simply run the following command:
pip install iwisdmIf you would like to install the package from source, you can clone the repository and follow the instructions below:
curl -sSL https://install.python-poetry.org | python3 -conda create --name iwisdm python=3.11poetry installTo initialize the ShapeNet environment, you will need to download the ShapeNet dataset, this is for rendering the trials.
To replicate our experiments, you also need to download the benchmarking configurations.
ShapeNet is a large-scale repository of shapes represented by 3D CAD models of objects (Chang et. al. 2015).
# imports
from iwisdm import make
from iwisdm import read_write
# environment initialization
with open('your/path/to/env_config', 'r') as f:
config = json.load(f) # using pre-defined AutoTask configuration
env = make(
env_id='ShapeNet',
dataset_fp='your/path/to/shapenet_handpicked',
)
env.set_env_spec(
env.init_env_spec(
auto_gen_config=config,
)
)
# AutoTask procedural task generation and saving trial
tasks = env.generate_tasks(10) # generate 10 random task graphs and tasks
_, (_, temporal_task) = tasks[0]
trials = env.generate_trials(tasks=[temporal_task]) # generate a trial
imgs, _, info_dict = trials[0]
read_write.write_trial(imgs, info_dict, f'output/trial_{i}')See /tutorials for more examples.
This repository builds upon the foundational work presented in the COG paper (Yang et al.).
Yang, Guangyu Robert, et al. "A dataset and architecture for visual reasoning with a working memory." Proceedings of the European Conference on Computer Vision (ECCV). 2018.
If you find iWISDM useful in your research, please use the following BibTex:
@InProceedings{pmlr-v274-lei25a,
title = {iWISDM: Assessing instruction following in multimodal models at scale},
author = {Lei, Xiaoxuan and Gomez, Lucas and Bai, Hao Yuan and Bashivan, Pouya},
booktitle = {Proceedings of The 3rd Conference on Lifelong Learning Agents},
pages = {457--480},
year = {2025},
editor = {Lomonaco, Vincenzo and Melacci, Stefano and Tuytelaars, Tinne and Chandar, Sarath and Pascanu, Razvan},
volume = {274},
series = {Proceedings of Machine Learning Research},
month = {29 Jul--01 Aug},
publisher = {PMLR},
pdf = {https://raw.githubusercontent.com/mlresearch/v274/main/assets/lei25a/lei25a.pdf},
url = {https://proceedings.mlr.press/v274/lei25a.html}
}

