\

Why are your models so big? (2023)

13 points - last Tuesday at 10:25 PM

Source
  • siddboots

    yesterday at 11:55 PM

    I think I have almost the opposite intuition. The fact that attention models are capable of making sophisticated logical constructions within a recursive grammar, even for a simple DSL like SQL, is kind of surprising. I think it’s likely that this property does depend on training on a very large and more general corpus, and hence demands the full parameter space that we need for conversational writing.