Jinn's Hub
about / blog / projects / EN /
所有标签

标签: "attention"

    注意力机制详解 — Full, Sparse, Linear, NSA & GLA
    从 Full Attention 出发,拆解 Sparse 和 Linear 两条路线,直到 DeepSeek NSA 和 Gated Linear Attention
© 2026 • Jinn's Hub 🔬
Press Esc or click anywhere to close