
DeepSeek, a Chinese AI startup, has developed the R1 model, which rivals leading AI reasoning models such as OpenAI’s O1. Using a Mixture-of-Experts (MoE) architecture, R1 activates only 37 billion of its 671 billion parameters during processing, significantly reducing computational costs and energy consumption while maintaining top-tier performance
Recent Comments