Skip to the content.

Literature

Background on LLMs, GPTs, and transformers

Our original inspiration: “Revealing the mystery behind chain of thought”

Other CoT analyses

suggested by Aziz

Elementary properties of the attention operator

Connections between LLMs and biology

suggested by Aziz

Tackling simple arithmetic using GPT architectures

Representation of concepts within LLMS

suggested by Aziz

The effect of prompts

suggested by Aziz

Augmenting transformer with memory

Tool use

suggested by Biruk

Alternatives to attention

Super-weights

Credit: Hemanth Kepa