Looks like a math improvement? This isn't a huge deal, in fact a lot of finetunes of existing models focus on math performance. InternLM just released some really interesting ones.
Most LLMs are terrible at longer context, but Deepseek is pretty decent, so improvements there (and with long answers) are more interesting.
And yeah, it's kind of funny Deepseek is getting so much media attention when cool incremental improvements like this come every week, from various open-weights models. It's awesome that they are releasing the weights, but still.