-
Notifications
You must be signed in to change notification settings - Fork 4
Open
Description
Hi, I'm trying to test my own model on DevBench and look at the difference between model prediction & human prediction. I'm able to generate the embedding features and prediction logits by running eval.py, but I'm having some troubles while parsing the human prediction data. Could you please help me with this? The details are listed below:
- For the TROG task, the number of trials in this codebase (N=78) is far less than what you mentioned in the paper (N=514), is there data missing or is the rest of the data unable to be released due to privacy constraints?
- For the Viz Vocab task, there are 119 testing samples in the
manifest.csvthat go through the model, but only 108 human trials data in your providedhuman.csv; neither is consistent with what you wrote in the paper (N=1780). Is there data missing or am I getting something wrong? - For the THINGS task, again the number of samples (N=1854) is inconsistent with the number in your paper (N=12340), but is consistent with the original paper revealing interpretable object representations from human behavior. May I know how you calculated the number and samples for this task?
- For the VOC task, it seems that the human data is stored in the
human.rdsfile. I guess it's an R file and tried to parse it using some python package, but failed to do so. Since there are many people who are familiar with python but not with R, is it possible for you to provide an alternative file that is readable by python? I'll really appreciate that! - For the WAT task, I'm totally lost in retrieving human data. I can see there's an
entwisle_norms.csvfile and severalCue_Target_Pairsfiles, but I have no clue what they mean. Could you please elaborate on the format of these human performance data file, or provide a code template that is able to parse the human data?
Thank you very much for your patience, and I appreciate any help you can provide!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels