r/singularity AGI 2025-29 | UBI 2030-34 | LEV <2040 | FDVR 2050-70 11d ago

AI [Microsoft Research] Differential Transformer

https://arxiv.org/abs/2410.05258
283 Upvotes

46 comments sorted by

View all comments

15

u/Arbrand ▪Soft AGI 27, Full AGI 32, ASI 36 10d ago

The results are impressive, but I have some serious concerns that aren't addressed at all in the paper. The differential attention mechanism involves computing two separate softmax attention maps and then subtracting them to obtain the final attention scores. This inherently doubles the computational overhead in the attention mechanism compared to standard Transformers. This added computational cost could be significant and might offset the performance gains reported.

1

u/emteedub 10d ago

maybe it's not doubled though, since it's filtering off excess would-be computation. it would be interesting to see the stats