News

A newly released 14-page technical paper from the team behind DeepSeek-V3, with DeepSeek CEO Wenfeng Liang as a co-author, sheds light on the “Scaling Challenges and Reflections on Hardware for AI ...
DeepSeek AI has announced the release of DeepSeek-Prover-V2, a groundbreaking open-source large language model specifically designed for formal theorem proving within the Lean 4 environment. This ...
The remarkable success of OpenAI’s o1 series and DeepSeek-R1 has unequivocally demonstrated the power of large-scale reinforcement learning (RL) in eliciting sophisticated reasoning behaviors and ...
Beijing, China – April 15, 2025 – In a strategic move that underscores its technological prowess and global ambitions, potentially paving the way for a future IPO, Chinese AI company Zhipu.AI has ...
On April 15, self-driving truck company TuSimple raised USD 1.35 billion with its NASDAQ IPO, with company valuation hitting USD 8.5 billion at high point. Founded in 2015, TuSimple had received ...
DeepSeek AI, a prominent player in the large language model arena, has recently published a research paper detailing a new technique aimed at enhancing the scalability of general reward models (GRMs) ...
The quality and fluency of AI bots’ natural language generation are unquestionable, but how well can such agents mimic other human behaviours? Researchers and practitioners have long considered the ...
A pair of groundbreaking research initiatives from Meta AI in late 2024 is challenging the fundamental “next-token prediction” paradigm that underpins most of today’s large language models (LLMs). The ...
The Transformer architecture, introduced by Vaswani et al. in 2017, serves as the backbone of contemporary language models. Over the years, numerous modifications to this architecture have been ...
Large Language Models (LLMs) have become indispensable tools for diverse natural language processing (NLP) tasks. Traditional LLMs operate at the token level, generating output one word or subword at ...
Researchers from Google DeepMind introduce the concept of "Socratic learning." This refers to a form of recursive self-improvement in artificial intelligence that significantly enhances performance ...
For artificial intelligence to thrive in a complex, constantly evolving world, it must overcome significant challenges: limited data quality and scale, and a lag in new, relevant information creation.