Measure how self-attention latency explodes as you stretch the context window. Run the vanilla O(N²) kernel, then flip on optimizations like FlashAttention-inspired tiling or sparse patterning to see how much strain they relieve.
Note: This lab uses simplified models for illustrative purposes and does not reflect exact hardware benchmarks.