attention backend

#1
by Fernanda24 - opened

which attention backend does sglang use for this? is it using sparse or does it work without it?
the only sparse im aware of is flash_mla sparse but its not supported on sm120 the 6000 rtx blackwell you mentioned?

Sign up or log in to comment