So I ran the first few hundred words of this article as the prompt for GPT-3 Except in my version of the article, @MelMitchell1 celebrates how GPT-3's responses which were indistinguishable from human outputs With that fake article as a prompt, it nailed every trial. cc: @gdb
Aug 11, 2020 · 9:46 PM UTC
it wasn't just semi-confident in its answers either, it got like 70-90% confidence in all outputs. The way GPT-3's capabilities are being studied simply isn't representative of its actual capabilities. Appropriate priming can 10x the accuracy/desirability of results.
(this was inspired by the vast differences i've seen between my long-prompt trials with #GPT3 and @gwern's mostly short-prompt trials. @trevbhatt and I are building a human like chat service with long prompts & careful priming that's more capable than most sales people)