How are current large language models being judged with apparent confidence, without a ranking of brain parameter size versus task performance metrics?

⚫Question for Prof Gary Marcus /Gary Marcus re biological brain parametrization size/current LLM sizes wrt task performance:

Q-A: Since general intelligence occurs naturally in around 1000 trillion parameters/synaptic connections (or in a supposedly rare case, 10% of that, in case where a man lost 90% of brain volume, due to hydrocephalus, but still reportedly functioned normally — Feuillet et al/The Lancet)…:

1. Would you expect today’s models, with far less parameters, to exhibit general intelligence?

2. If not what degree of general intelligence is expected given the current/apparent parameter mis-match so far?

3. How do you determine whether the degree seen in ChatGPT for eg, isn’t consistent with some expected measure, given the parameter mis-match?

Q-B: Is there a ranking somewhere of expected task performance wrt ranges of neuro-computational parameter size requirements?

If not, how are judgments being cast?

Question maker: G. Bennett

--

--

Lecturer of Artificial Intelligence, and inventor of “Supersymmetric Deep Learning” → Github/Supersymmetric-artificial-neural-network

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
God Bennett

Lecturer of Artificial Intelligence, and inventor of “Supersymmetric Deep Learning” → Github/Supersymmetric-artificial-neural-network