Why are your models so big? (2023)
13 points - last Tuesday at 10:25 PM
Sourcesiddboots
yesterday at 11:55 PM
I think I have almost the opposite intuition. The fact that attention models are capable of making sophisticated logical constructions within a recursive grammar, even for a simple DSL like SQL, is kind of surprising. I think itβs likely that this property does depend on training on a very large and more general corpus, and hence demands the full parameter space that we need for conversational writing.