Cornell University
Library
Cornell UniversityLibrary

eCommons

Help
Log In(current)
  1. Home
  2. Cornell University Graduate School
  3. Cornell Theses and Dissertations
  4. Memory in Language Models: Representation and Extraction

Memory in Language Models: Representation and Extraction

File(s)
Morris_cornellgrad_0058F_15353.pdf (3.87 MB)
Permanent Link(s)
https://doi.org/10.7298/rgq1-w286
https://hdl.handle.net/1813/121141
Collections
Cornell Theses and Dissertations
Author
Morris, John
Abstract

We explore memory in neural language models as it is stored in model weights after training and in model activations ("embeddings") during inference. We first describe a method for improving embeddings by incorporating surrounding documents in the context of text retrieval. We then propose a new method for inverting text embeddings and demonstrate its applicability to the outputs of general language models. Finally, we measurement of the information content of model weights, a way to characterize the total amount of information models can store.

Description
138 pages
Date Issued
2025-12
Keywords
artificial intelligence
•
embeddings
•
information theory
•
inversion
•
language models
•
LLMs
Committee Chair
Rush, Alexander
Committee Member
Zabih, Ramin
Pierson, Emma
Degree Discipline
Computer Science
Degree Name
Ph. D., Computer Science
Degree Level
Doctor of Philosophy
Rights
Attribution 4.0 International
Rights URI
https://creativecommons.org/licenses/by/4.0/
Type
dissertation or thesis

Site Statistics | Help

About eCommons | Policies | Terms of use | Contact Us

copyright © 2002-2026 Cornell University Library | Privacy | Web Accessibility Assistance