A markdown doc or something would suffice -- info with possible code snippets on how to launch an on-demand evaluation, how to turn on/off live evals, what metrics are available and what they mean, how to download results and their format so people can see how to display it in a UI should they want to do so, etc.
Eventually if a fullstack sample is published we can link to it from this doc as well.