I think embodiment and the encoding of the natural laws (gravity, force, etc) that go into that will be another huge step at grounding AI. People tend to gravitate to thinking about humanoid robots when that is mentioned (and thereby terminators), but honestly I would think things closer to sensor networks involving thousands or millions of bodies like a hivemind would be more likely (why stick at the human level of a single body if you didn't have to). Interaction with the world is a means of determining truth... The ability to perform science.
And as hard as embodiment is, it will be the easy part in my opinion. Continuous learning without losing the plot is going to be quite the challenge. If an LLM has something wrong, how does it update and change that bit of information without huge amounts of power use? How do you make the system learn 'important' things without filling up with junk/spam it is subject to? How do you keep the system aligned with a goal that is not destructive to itself or others?
But embodiment being a bottleneck could indicate that it's a data/training issue, rather than an architectural issue. Multimodal training data improves GPT-4 already, but that's still very little data compared to growing up to a full human adult. There are still many things to try.
I think embodiment and the encoding of the natural laws (gravity, force, etc) that go into that will be another huge step at grounding AI. People tend to gravitate to thinking about humanoid robots when that is mentioned (and thereby terminators), but honestly I would think things closer to sensor networks involving thousands or millions of bodies like a hivemind would be more likely (why stick at the human level of a single body if you didn't have to). Interaction with the world is a means of determining truth... The ability to perform science.
And as hard as embodiment is, it will be the easy part in my opinion. Continuous learning without losing the plot is going to be quite the challenge. If an LLM has something wrong, how does it update and change that bit of information without huge amounts of power use? How do you make the system learn 'important' things without filling up with junk/spam it is subject to? How do you keep the system aligned with a goal that is not destructive to itself or others?