Discussion about this post

User's avatar
Tinah MB's avatar

I might take burritos a bit too seriously, but I think the metaphor actually holds up, the ingredients go together for a reason. That’s kind of the whole “vision informs memory” thing: each module adds something that changes how the others taste.

Expand full comment
Tim Duffy's avatar

One thing I've been wondering about is whether normal LLMs could satisfy the GWT indicator properties by default. Some half-formed thoughts from thinking aloud:

GWT-1: There are in fact two places where individual modules are used in LLMs, experts in MoEs and attention heads. You could think of one or both of these as modules, and the residual stream as the workspace. However I think the analogy is not great. In attention heads, there's no competing for global representation, the outputs are just concatenated and then added to the residual stream. For experts, the expert selection is done based on expected relevance, but this is before the modules have a chance to do their thinking instead of after.

GTT-2: As I speculated above, I think the residual stream is one candidate for being a workspace. But its case seems weak. Layer-by-layer updates to the residual stream are incremental, and so the focus doesn't change dramatically by layer based on the module selected. It changes a lot from token to token, but that's after being discarded after the end of each one.

GWT-3: Both experts and attention heads satisfy this one, they both contribute to the residual stream, and take its current value as input when they're used.

GWT-4: TBH not sure how to evaluate this one. There's not really much flashy new information while a token is being processed.

Overall this doesn't make it seem like the conditions are satisfied, although there is information that is locally computed and globally used.

Expand full comment
1 more comment...

No posts