attention backend
#1
by
Fernanda24
- opened
which attention backend does sglang use for this? is it using sparse or does it work without it?
the only sparse im aware of is flash_mla sparse but its not supported on sm120 the 6000 rtx blackwell you mentioned?