As an experiment, I was considering fine-tuning an LLM (models like 'ChatGPT') on WOBs. They generally fit the training format of question: answer that makes for a good, and in this case, fun training dataset. While I could probably do web scraping and eventually pull in all the text, with the website itself open source I figured there must be a copy of the database somewhere. Anyone know where I could find this?