This is a repo to organize the official ipynb implementation of SayCan (Do As I Can, Not As I Say: Grounding Language in Robotic Affordances) for easier further research.
Clone this repo. Create and activate new conda environment with python 3.9. Run the following command.
pip install -r requirements.txt
gdown --id 1Cc_fDSBL6QiDvNT4dpfAEbhbALSVoWcc
gdown --id 1yOMEm-Zp_DL3nItG9RozPeJAmeOldekX
gdown --id 1GsqNLhEl9dd4Mc3BM0dX3MibOI1FVWNM
unzip ur5e.zip
unzip robotiq_2f_85.zip
unzip bowl.zip
gsutil cp -r gs://cloud-tpu-checkpoints/detection/projects/vild/colab/image_path_v2 ./
You can skip this process if you want to generate data by yourself with gen_data.py.
Download pregenerated dataset by running
gdown --id 1yCz6C-6eLWb4SFYKdkM-wz5tlMjbG2h8
gdown --id 1Nq0q1KbqHOA5O7aRSu4u7-u27EMMXqgP
Don't forget to add your openai key in llm.py.
If you have downloaded the pretrained policy in 2.4, you can now run demo.py to visualize the evaluation process.
If you want to train a model from scratch, run train.py.