Blog

  • DeepSeek has released its source code, detailed explanation of FlashMLA
    Last week, DeepSeek announced that it would open source five projects next week: Netizens said, “This time, OpenAI is really here.” Just now, the first open source project came, related to inference acceleration, FlashMLA: Open source project address: DeepSeek FlashMLA It has been open source for two hours, and Github already has 2.7k+ stars: The…
  • What is FlashMLA? A Comprehensive Guide to Its Impact on AI Decoding Kernels
    FlashMLA has quickly gained attention in the world of artificial intelligence, particularly in the field of large language models (LLMs). This innovative tool, developed by DeepSeek, serves as an optimized decoding kernel designed for Hopper GPUs—high-performance chips commonly used in AI computations. FlashMLA focuses on the efficient processing of variable-length sequences, making it particularly well-suited…
  • Qwen2.5-max vs DeepSeek R1: A deep comparison of models: a full analysis of application scenarios
    Introduction Today, large language models (LLMs) play a crucial role. In early 2025, as the competition for AI intensified, Alibaba launched the new Qwen2.5-max AI model, and DeepSeek, a company from Hangzhou, China, launched the R1 model, which represents the pinnacle of LLM technology. Deepseek R1 is an open source AI model that has attracted…
  • It is close to DeepSeek-R1-32B and crushes Fei-Fei Li’s s1! UC Berkeley and other open source new SOTA inference models
    The 32B inference model uses only 1/8 of the data and is tied with DeepSeek-R1 of the same size! Just now, institutions such as Stanford, UC Berkeley, and the University of Washington have jointly released an SOTA-level inference model, OpenThinker-32B, and have also open-sourced up to 114k training data. OpenThinker Project homepage: OpenThinker Hugging Face:…

Create your free account today!