LongCat-Flash-Thinking, LLM from Meituan (China's Equivalent of Uber Eats)
github.comNo clue why they are calling a 560B model "Flash". But one very interesting thing is, it is beating all the other frontier models in safety, and most of them by quite a margin. The same for Formal Theorem Proving, where I would have expected OpenAI to do way better