🚨The DeepSeek team cracked cheap long context for LLMs: a ~3.5x cheaper prefill and ~10x cheaper decode at 128k context at inference with the same quality. Can’t help but love the only frontier AI lab that publishes an open model and great open research!
Tech Report:
62.46K