David Mimno is an Associate Professor and Chair of the Department of Information Science in the Ann S. Bowers College of Computing and Information Science at Cornell University. He holds a Ph.D. from UMass Amherst and was previously the head programmer at the Perseus Project at Tufts as well as a researcher at Princeton University. Professor Mimno’s work has been supported by the Sloan Foundation, the NEH, and the NSF.
LLM Architectures and EmbeddingsCornell Course
Course Overview
In this course, you will investigate the internal workings of transformer-based language models by exploring how embeddings, attention, and model architecture shape textual outputs. You'll begin by building a neural search engine that retrieves documents through vector similarity then move on to extracting token-level representations and visualizing attention patterns across different layers and heads.
As you progress, you will analyze how tokens interact with each other in a large language model (LLM), compare encoder-based architecture with decoder-based architectures, and trace how a single word's meaning can shift from input to output. By mastering techniques like plotting similarity matrices and identifying key influencers in the attention process, you'll gain insights enabling you to decode model behaviors and apply advanced strategies for more accurate, context-aware text generation.
You are required to have completed the following courses or have equivalent experience before taking this course:
- LLM Tools, Platforms, and Prompts
- Language Models and Next-Word Pronunciation
- Fine-Tuning LLMs
- Language Models and Language Data
Key Course Takeaways
- Identify the tokens a model focuses on within an input sentence, analyzing how each token influences or is influenced by others
- Extract and visualize attention matrices for sample text
- Compare various LLM architectures to understand how token interactions differ across encoder, decoder, and encoder-decoder models
- Examine intermediate token representations in upper and middle layers, using vector similarity to reveal context-specific shifts in meaning
- Implement and evaluate document retrieval based on embedding vectors, showcasing how learned representations enable effective content search

How It Works
Course Author
Who Should Enroll
- Engineers
- Developers
- Analysts
- Data scientists
- AI engineers
- Entrepreneurs
- Data journalists
- Product managers
- Researchers
- Policymakers
- Legal professionals
100% Online
cornell's Top Minds
career