>>55157376Not necessarily "controled", it's more just different "parts" that learn to do something "specialized". It also bumps up the total number of parameters by a lot: I believe the total amount for GPT-4 is 1.8 TRILLION.
And here I was laughing at that rumor that GPT-4 would have one trillion parameters. They actually did it.