DeepSeek announced a new experimental AI model featuring Sparse Attention technology that dramatically lowers inference costs for long‑context tasks. The model, released on Hugging Face and accompanied by a research paper on GitHub, uses a lightning indexer and fine‑grained token selection to focus computational resources on the most relevant excerpts. Preliminary tests suggest API call prices can be cut by as much as 50 percent in long‑context scenarios. The open‑weight release invites third‑party validation and positions DeepSeek as a notable player in the ongoing effort to make transformer‑based AI more cost‑effective.
Leer más →