Skip to main page content

Day 10 | Large Language Models & Multimodal Grounding

My Session Status

When:
9:00 AM, Friday 14 Jun 2024 EDT (8 hours)
Theme:
Large Language Models & Multimodal Grounding

Sub Sessions

9:00 AM EDT - 10:30 AM EDT | 1 hour 30 minutes
Large Language Models & Multimodal Grounding

Do LLMs understand the meaning of the texts they generate? Do they possess a semantic grounding? And how could we understand whether and what they understand? We have recently witnessed a generative turn in AI, since generative models, including LLMs, are key for self-supervised learning. To assess the question of semantic grounding, I distinguish and discuss five methodological ways. The most promising way is to apply core assumptions of theories of meaning in philosophy of mind and language...

11:00 AM EDT - 12:30 PM EDT | 1 hour 30 minutes
Large Language Models & Multimodal Grounding

Large language models (LLMs) appear to exhibit human-level abilities on a range of tasks, yet they are notoriously considered to be “black boxes”, and little is known about the internal representations and mechanisms that underlie their behavior. This talk will discuss recent work which seeks to illuminate the processing that takes place under the hood. I will focus in particular on questions related to LLM’s ability to represent abstract, compositional, and content-independent operations of ...

1:30 PM EDT - 3:00 PM EDT | 1 hour 30 minutes
Large Language Models & Multimodal Grounding

Over the last decade, multimodal vision-language (VL) research has seen impressive progress. We can now automatically caption images in natural language, answer natural language questions about images, retrieve images using complex natural language queries and even generate images given natural language descriptions.Despite such tremendous progress, current VL research faces several challenges that limit the applicability of state-of-art VL systems. Even large VL systems based on multimodal l...

3:30 PM EDT - 5:00 PM EDT | 1 hour 30 minutes
Large Language Models & Understanding

Large language models have been found to have surprising capabilities, even what have been called “sparks of artificial general intelligence.” However, understanding these models involves some significant challenges: their internal structure is extremely complicated, their training data is often opaque, and getting access to the underlying mechanisms is becoming increasingly difficult. As a consequence, researchers often have to resort to studying these systems based on their behavior. This s...

My Session Status

Session detail
Allows attendees to send short textual feedback to the organizer for a session. This is only sent to the organizer and not the speakers.
To respect data privacy rules, this option only displays profiles of attendees who have chosen to share their profile information publicly.

Changes here will affect all session detail pages