Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The real world is governed by physics, so isn't "next state prediction" a sufficient eval function that forces it to internalize a world model? And increasing the timespan over which you predict requires an increasing amount of "intelligence" because it requires modeling the real-world behavior of constituent subsystems that are often black-boxes (e.g. if there is a crow on a road with a car approaching, you can't just treat it as a physics simulation, you need to know that crows are capable of flying to understand what is going to happen).

I don't see how this is any less structured than the CLM objective of LLMs, there's a bunch of rich information there.



If Transformers were able to learn in such a trivial fashion, we would already have AGI.

There is at least one missing piece to the puzzle, and some say 5-6 more breakthrough are necessary.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: