Last week, DeepSeek announced that it would open source five projects next week: Netizens said, “This time, OpenAI is really here.” Just now, the first open source project came, related to inference acceleration, FlashMLA: Open source project address: DeepSeek FlashMLA It has been open source for two hours, and Github already has 2.7k+ stars: The…
FlashMLA has quickly gained attention in the world of artificial intelligence, particularly in the field of large language models (LLMs). This innovative tool, developed by DeepSeek, serves as an optimized decoding kernel designed for Hopper GPUs—high-performance chips commonly used in AI computations. FlashMLA focuses on the efficient processing of variable-length sequences, making it particularly well-suited…
Introduction Today, large language models (LLMs) play a crucial role. In early 2025, as the competition for AI intensified, Alibaba launched the new Qwen2.5-max AI model, and DeepSeek, a company from Hangzhou, China, launched the R1 model, which represents the pinnacle of LLM technology. Deepseek R1 is an open source AI model that has attracted…
The 32B inference model uses only 1/8 of the data and is tied with DeepSeek-R1 of the same size! Just now, institutions such as Stanford, UC Berkeley, and the University of Washington have jointly released an SOTA-level inference model, OpenThinker-32B, and have also open-sourced up to 114k training data. OpenThinker Project homepage: OpenThinker Hugging Face:…