Join us on Wednesday, October 1st, at 5 pm (CEST) for the online event “Multimodal LLMs: AI that Can Read, Look, Listen, and Speak ” on the MS Teams platform.
This talk is an accessible, exploratory tour of Multimodal Large Language Models (MLLMs), AI systems that can understand and generate across text, images, audio, video, and more.
We’ll unpack what “multimodal” really means, how these models differ from text-only LLMs, and why this shift represents a new paradigm in AI. Through intuitive examples and live-style walkthroughs (no coding or prior expertise required), you’ll see how MLLMs can describe images, read charts and documents, follow spoken prompts, and connect ideas across different kinds of media.
The goal is awareness and curiosity: to give you a clear mental model of what’s now possible, common terms you’ll encounter, and a sense of where the field is heading, so you can recognize and engage with this emerging wave of AI.
Moderator:
Claudio Sartori | Director of the Master in Data Science and Business Analytics
The event will be held in English.
To participate, please fill out the registration form.