Playing chess with large language models:
I built a bot to play chess by querying a text language model. It sees the sequence of moves in order (as text!), and predicts which move comes next. It's better than me.
A GPT-4 Forecasting Challenge:
Test your ability to predict (in a calibrated manner) whether or not GPT-4 can answer a range of questions from coding to poetry to baking.
A ChatGPT clone, in 3000 bytes of C, backed by GPT-2:
A dependency-free implementation of GPT-2, including byte-pair encoding and transformer inference, in ~3000 bytes of C. I then use this to create something like Chat GPT.