Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Cognitive benchmarking in large models

Overview

In this session, participants will explore the fundamentals of Large Language Models (LLMs), Vision-Language Models (VLMs), and reasoning models. The LLM and VLM sections will each include:

  1. A concise theoretical overview

  2. Key practical considerations

  3. Hands-on exercises

Instructors

Lucas Gomez is a PhD student in the Integrated Program in Neuroscience at McGill University. His research in the BashivanLab focuses on building large scale deep learning models of visual working memory that predict prefrontal cortex activity. He has a Bachelors of Computer Science from Carleton University in Ottawa, and he is broadly interested in the intersection between AI and the brain sciences.

Zihan Weng is a Ph.D. student in the Integrated Program in Neuroscience at McGill University, supervised by Pouya Bashivan, where he also completed his master’s degree. His research focuses on drawing inspiration from the brain’s hierarchical memory systems to develop more efficient and scalable deep learning architectures. He earned a master’s degree in Biomedical Engineering from UESTC and a bachelor’s degree in Computer Science. He is interested in how biological neural networks can help us build better artificial neural networks.

Declan Campbell is a PhD student in the Princeton Neuroscience Institute and a Natural and Artificial Minds (NAM) graduate fellow, advised by Jonathan Cohen and Tom Griffiths. His research draws on behavioral paradigms from cognitive psychology and methods from mechanistic interpretability to study the mechanisms underlying abstract reasoning in humans and machines.

Objectives

This session is divided into three parts:

Large Language Models (LLMs)

Vision-Language Models (VLMs)

Reasoning Models

Materials