{"id":677,"date":"2024-11-06T12:00:06","date_gmt":"2024-11-06T12:00:06","guid":{"rendered":"https:\/\/projects.illc.uva.nl\/indeep\/?page_id=677"},"modified":"2024-12-09T08:49:42","modified_gmt":"2024-12-09T08:49:42","slug":"indeep-video-series","status":"publish","type":"page","link":"https:\/\/projects.illc.uva.nl\/indeep\/indeep-video-series\/","title":{"rendered":"InDeep Video Series"},"content":{"rendered":"\n<h2 class=\"wp-block-heading has-text-align-center\">Six videos that explain and explore <br>how to open the Black Box <br>of Large Language Models<\/h2>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"389\" src=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-1024x389.png\" alt=\"\" class=\"wp-image-701\" style=\"width:905px;height:auto\" srcset=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-1024x389.png 1024w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-300x114.png 300w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-768x292.png 768w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-1536x584.png 1536w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/Blackboxtitle2-2048x778.png 2048w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n<div class=\"wp-block-group is-nowrap is-layout-flex wp-container-core-group-is-layout-ad2f72ca wp-block-group-is-layout-flex\">\n<p class=\"has-medium-font-size\">With the enormous impact that Artificial Intelligence models have in science, technology and society, many people are looking for ways to make these models more transparent to better understand what is happening on the inside of these models. In the field of Explainable AI researchers have come up with many different tools to help achieve these goals. These tools are called \u2018interpretability methods\u2019. Understanding the limits of interpretability methods, as well as finding ways to understand more of what is happening inside the black box of Transformer models is therefore crucially important, for both engineers and policy makers.&nbsp;<br><br>In this video series made by the InDeep research group and presented by our (affiliated) researchers we will answer why Interpretability is important in the Age of LLMs, explain the behaviour of neural models, demonstrate why it is crucial to track how Transformers mix contextual information, show how to best measure context-mixing in Transformers, what a circuit is, and how to use circuits to reverse-engineer model mechanisms.<\/p>\n\n\n\n<p class=\"has-medium-font-size\"><br><\/p>\n<\/div>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"285\" src=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s-1024x285.png\" alt=\"\" class=\"wp-image-750\" style=\"width:895px;height:auto\" srcset=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s-1024x285.png 1024w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s-300x83.png 300w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s-768x214.png 768w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s-1536x427.png 1536w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/12-3s.png 1599w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n<p class=\"has-medium-font-size\"><br>In the <a href=\"https:\/\/youtu.be\/k3C7NRpu0ss?si=U_tEytvMDIu19FxR\" target=\"_blank\" rel=\"noreferrer noopener\">first video<\/a> of this series we will explore the state of the field and give an introduction to how to begin the extremely challenging task to explain the behaviour of <em>large-scale neural networks<\/em>. In our <a href=\"https:\/\/youtu.be\/H-EOrI520uU?si=cb2QBc32uXrtgUK2\" target=\"_blank\" rel=\"noreferrer noopener\">second video<\/a> we look at how we can leverage insights from the field of cognitive science to help us interpret large-scale neural networks. However, as this video also shows, merely employing general-purpose interpretability methods may not lead to reliable findings.\u00a0In order to resolve this lack, we may need to interpret models at a more fine-grained level.<br><br>1- <a href=\"https:\/\/youtu.be\/k3C7NRpu0ss?si=U_tEytvMDIu19FxR\" target=\"_blank\" rel=\"noreferrer noopener\">Why Interpretability is important in the A<\/a><a href=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video1a_final_small.mp4\">ge of LLMs<\/a> Dr Jelle Zuidema, Associate professor, University of Amsterdam<br>2- <a href=\"https:\/\/youtu.be\/H-EOrI520uU?si=cb2QBc32uXrtgUK2\" target=\"_blank\" rel=\"noreferrer noopener\">How can we explain the behaviour of neural models?<\/a> Jaap Jumelet, PhD candidate, University of Amsterdam<br><\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"283\" src=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2-1024x283.png\" alt=\"\" class=\"wp-image-748\" style=\"width:888px;height:auto\" srcset=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2-1024x283.png 1024w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2-300x83.png 300w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2-768x212.png 768w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2-1536x425.png 1536w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/34v2-2.png 1952w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n<p class=\"has-medium-font-size\"><br>In the next two videos we dive deeper into the inner working of interpretability methods and demonstrate various methods that have been proposed to create explanations that are more faithful to a model\u2019s actual behaviour. The <a href=\"https:\/\/youtu.be\/8Nj1xiV-hA0?si=GGt6w9Xo0QRW2xz1\" target=\"_blank\" rel=\"noreferrer noopener\">third video<\/a> introduces <em>context mixing<\/em> and discuss its importance for understanding how <em>Transformers <\/em>process data and build contextualized representations. In <a href=\"https:\/\/youtu.be\/JPOBPY-ndfk?si=0aqq5o-7O6xKx1Cr\" target=\"_blank\" rel=\"noreferrer noopener\">video four<\/a> we will learn about methods to measure context-mixing in Transformers but also that some of these methods are not a well-equipped as they first seem.<br><br>3- <a href=\"https:\/\/youtu.be\/8Nj1xiV-hA0?si=GGt6w9Xo0QRW2xz1\" target=\"_blank\" rel=\"noreferrer noopener\">Why it is crucial to track how Transformers mix contextual information<\/a> Dr Afra Alishahi, Associate professor, Tilburg University<br>4- <a href=\"https:\/\/youtu.be\/JPOBPY-ndfk?si=0aqq5o-7O6xKx1Cr\" target=\"_blank\" rel=\"noreferrer noopener\">How to best measure context-mixing in Transformers<\/a> Hosein Mohebbi, PhD candidate, Tilburg University<\/p>\n\n\n<div class=\"wp-block-image\">\n<figure class=\"aligncenter size-large is-resized\"><img loading=\"lazy\" decoding=\"async\" width=\"1024\" height=\"284\" src=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s-1024x284.png\" alt=\"\" class=\"wp-image-706\" style=\"width:897px;height:auto\" srcset=\"https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s-1024x284.png 1024w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s-300x83.png 300w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s-768x213.png 768w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s-1536x426.png 1536w, https:\/\/projects.illc.uva.nl\/indeep\/wp-content\/uploads\/2024\/11\/video56-s.png 1949w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/figure><\/div>\n\n\n<p class=\"has-medium-font-size\"><br>In the final two videos of this series, we&#8217;ll explore other types of Transformer-specific methods that involve the model&#8217;s decision and find subnetworks in the model that are responsible for performing specific subtasks. In <a href=\"https:\/\/youtu.be\/JfukUIB3JDE?si=qwf9z7O-wvCHVNBZ\" target=\"_blank\" rel=\"noreferrer noopener\">video five<\/a> we will introduce <em>circuits<\/em>, a new framework for providing low-level, algorithmic explanations of language model behavior. But how do you find a circuit? In the <a href=\"https:\/\/youtu.be\/ubZb_gE9638?si=0caPK3I3l25R_QrY\" target=\"_blank\" rel=\"noreferrer noopener\">sixth video<\/a> of this series, we\u2019ll explain the techniques used to uncover circuits for any task.<\/p>\n\n\n\n<p class=\"has-medium-font-size\">5- <a href=\"https:\/\/youtu.be\/JfukUIB3JDE?si=qwf9z7O-wvCHVNBZ\" target=\"_blank\" rel=\"noreferrer noopener\">What is a circuit, and how does it explain LLM behavior?<\/a> Dr Sandro Pezzelle, Assistant professor, University of Amsterdam<br>6- <a href=\"https:\/\/youtu.be\/ubZb_gE9638?si=0caPK3I3l25R_QrY\" target=\"_blank\" rel=\"noreferrer noopener\">How to reverse-engineer model mechanisms by finding circuits<\/a> Michael Hanna, PhD candidate, University of Amsterdam<\/p>\n\n\n\n<p class=\"has-medium-font-size\">We hope you will enjoy and learn from these first set of videos. We hope to make more video\u2019s as there\u2019s plenty more in the world of interpretability and circuits beyond this.\u00a0<\/p>\n\n\n\n<p class=\"has-medium-font-size\"><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Six videos that explain and explore how to open the Black Box of Large Language Models With the enormous impact that Artificial Intelligence models have in science, technology and society, many people are looking for ways to make these models more transparent to better understand what is happening on the inside of these models. In&hellip;&nbsp;<a href=\"https:\/\/projects.illc.uva.nl\/indeep\/indeep-video-series\/\" rel=\"bookmark\">Read More &raquo;<span class=\"screen-reader-text\">InDeep Video Series<\/span><\/a><\/p>\n","protected":false},"author":2,"featured_media":0,"parent":0,"menu_order":0,"comment_status":"closed","ping_status":"closed","template":"","meta":{"neve_meta_sidebar":"","neve_meta_container":"","neve_meta_enable_content_width":"","neve_meta_content_width":0,"neve_meta_title_alignment":"","neve_meta_author_avatar":"","neve_post_elements_order":"","neve_meta_disable_header":"","neve_meta_disable_footer":"","neve_meta_disable_title":"","footnotes":""},"class_list":["post-677","page","type-page","status-publish","hentry"],"_links":{"self":[{"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/pages\/677","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/comments?post=677"}],"version-history":[{"count":45,"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/pages\/677\/revisions"}],"predecessor-version":[{"id":785,"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/pages\/677\/revisions\/785"}],"wp:attachment":[{"href":"https:\/\/projects.illc.uva.nl\/indeep\/wp-json\/wp\/v2\/media?parent=677"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}