CS Fall Seminar Series: Stefano Soatto
Who can attend?
Stefano Soatto, a professor of computer science at the University of California, Los Angeles, and a vice president at Amazon Web Services, where he leads the artificial intelligence labs, will give a talk titled "Toward Foundational Models of Physical Scenes: From Large Language Models to World Models and Back" for the Department of Computer Science.
This is a hybrid event; to attend virtually, use the Zoom link.
Now that a significant fraction of human knowledge has been shared through the internet, scraped and squashed into the weights of large language models (LLMs), do we still need embodiment and interaction with the physical world to build representations? Is there a dichotomy between LLMs and "large world models"? What is the role of visual perception in learning such models? Can perceptual agents trained by passive observation learn world models suitable for control?
To begin tackling these questions, I will first address the issue of controllability of LLMs. LLMs are tochastic dynamical systems, for which the notion of controllability is well established: The state ("of mind") of an LLM can be trivially steered by a suitable choice of input given enough time and memory. However, the space of interest for control of an LLM is not that of words, but that of "meanings" expressible as sentences that a human could have spoken and would understand. Unfortunately, unlike controllability, the notions of meaning and understanding are not usually formalized in a way that is relatable to LLMs in use today.
I will propose a simplistic definition of meaning that reflects the functional characteristics of a trained LLM. I will show that a well-trained LLM establishes a topology in the space of meanings, represented by equivalence classes of trajectories of underlying dynamical model (LLM). Then, I will describe both necessary and sufficient conditions for controllability in such a space of meanings.
I will then highlight the relation between meanings induced by a trained LLM upon the set of sentences that could be uttered, and "physical scenes" underlying sets of images that could be observed. In particular, a physical scene can be defined uniquely and inferred as an abstract concept without the need for embodiment, a view aligned with J. Koenderink's characterization of images as "controlled hallucinations."
Last, I will show that popular models ostensibly used to represent the 3D scene (Neural Radiance Fields, or NeRFs) can at most represent the images on which they are trained, but not the underlying physical scene. However, composing a NeRF with a latent diffusion model or other inductively-trained generative model yields a viable representation of the physical scene. Such a model class, which can be learned through passive observations, is a first albeit rudimentary foundational model of physical scenes in the sense of being sufficient for any downstream inference task based on visual data.
Who can attend?