Alternate Histories and GPT-3
Riskgaming - En podkast av Lux Capital
Kategorier:
"GPT-3 was trained on is so large that the model contains a certain fraction of the actual complexity of the world. But how much is actually inside these models, implicitly embedded within these neural networks? I decided to test this and see if I could examine the GPT-3 model of the world through the use of counterfactuals. Specifically, I wanted to see if GPT-3 could productively unspool histories of the world if things were slightly different, such as if the outcome of a war were different or a historical figure hadn’t been born. I wanted to see how well it could write alternate histories." - Samuel Arbesman From Cabinet of Wonders newsletter by Samuel Arbesman Great tweet thread summarizing his post "Securities" podcast is produced and edited by Chris Gates