The new research could be a harbinger of the company's next big model release after the R1. Credit: DeepSeek Chinese AI company Deepseek has unveiled a new training method, Manifold-Constrained Hyper-Connections (mHC), which will make it possible to train large language models more efficiently and at lower cost, reports the South China Morning Post. The method is a further development of so-called Hyper-Connections, which was originally developed by Bytedance in 2024. That technology, in turn, builds on the classic ResNet architecture from Microsoft Research Asia. Deepseek says mHC provides more stable and scalable training without increasing computational costs, thanks to specific optimizations at the infrastructure level. The researchers have tested the technology on models with up to 27 billion parameters with positive results. According to experts cited by the South China Morning Post, the new method could be a foretaste of the next big model release from Deepseek. The company launched its high-profile R1 model on the occasion of Chinese New Year 2025. Generative AIArtificial Intelligence SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe