
Tech
Mar 27, 2026VentureBeat
IndexCache Speeds Long-Context AI Models by 1.82x
IndexCache, a novel sparse attention optimizer by Tsinghua University and Z.ai, dramatically accelerates long-context AI models. It cuts up to 75% redundant computation, delivering up to 1.82x faster inference and significant cost savings.
Read →