BLUE
Profile banner
TU
Ted Underwood 🦋
@tedunderwood.me
Uses machine learning to study literary imagination, and vice-versa. Likely to share news about AI & computational social science / ciência social computacional / 計算社会科学. Information Sciences and English, UIUC. Author of Distant Horizons (Chicago, 2019).
4.6k followers2.9k following8.5k posts
TUtedunderwood.me

Greg Kamradt spent $200 to test the new 128,000-token context length on GPT-4. Can it remember a fact buried in tens of thousands of words of prose? The answer: yes, up to about 64k tokens. Beyond that, it matters where you placed the fact in the prompt. link to thread: twitter.com/GregKamradt/...

A heat map of document retrieval accuracy, showing that in very long contexts it gets harder for the model to remember it read something, esp if the fact is placed in the first half of the prompt and not right at the start.
0

Profile banner
TU
Ted Underwood 🦋
@tedunderwood.me
Uses machine learning to study literary imagination, and vice-versa. Likely to share news about AI & computational social science / ciência social computacional / 計算社会科学. Information Sciences and English, UIUC. Author of Distant Horizons (Chicago, 2019).
4.6k followers2.9k following8.5k posts