Skip to main contentPlayground
An Experimental MoE LLM introducing DeepSeek Sparse Attention (DSA). It achieves dramatically lower long-context inference cost while maintaining V3.1’s high quality.
Key Features
- Sparse Attention (DSA): Utilizes a novel sparse attention mechanism to cut inference costs significantly.
- Cost Efficiency: Achieves near-parity with V3.1 quality at a fraction of the computational expense.
- Long-Context Efficiency: Optimized for cost-effective handling of extremely large documents and code files.