Category: Large Language Models
- Causal Inference Charges into the LLM Battlefield! Large Model Hallucination Terminator? ABCA Framework
- Paper Brief | Activating and Enhancing Causal Reasoning Capabilities of Large Language Models Using Conditional Statements (CL2025)
- What? RLVR Isn't Learning New Knowledge—It's Learning How to Use Knowledge for Reasoning!
- Karpathy Forms a Large Model 'Parliament': GPT-5.1, Gemini 3 Pro, and Others Become the Strongest Think Tank
- Under $8,000! Sina Weibo's 1.5B Small Model Surpasses Near-Trillion Parameter Models
- Microsoft Proposes GAD Framework: Open-Source Models Can Directly Distill Black-Box GPT-5
- Making LLMs Work Like a Company: Microsoft Turns 'Concurrent Thinking' into a Protocol, Higher Accuracy and 28% Reduction in Critical Path Latency
- Reinforcement Learning + Large Model Memory: Mem-α, Enabling Agents to "Learn How to Remember" for the First Time
- AI Visionary Fei-Fei Li's Extensive Article Ignites Silicon Valley: Large Language Models Are on the Wrong Path, Spatial Intelligence Is the Only Way to AGI
- Meta's Two Latest Agent Learning Papers Are Quite Interesting!
- Inoculation Prompting: Making Large Language Models "Misbehave" During Training to Improve Test-Time Alignment
- We Planted a Word in Claude's Mind, and It Began to "Rationalize"! Anthropic's Latest Research: AI Possesses Introspective Abilities!
- GPT models becoming more conservative? Stanford Manning team proposes Verbalized Sampling to make models "think a bit more"
- Abandoning Manual Annotation! Chinese Team Proposes Self-Evolution Algorithm for Multimodal Large Models
- Abandoning Fine-Tuning: Stanford Co-releases Agentic Context Engineering (ACE), Boosting Model Performance by 10% and Reducing Token Costs by 83%
- Just Released! Tsinghua and Partners Open Source UltraRAG 2.0! Performance Soars by 12%
- Google Enters the CUA Battleground, Launches Gemini 2.5 Computer Use: Allowing AI to Directly Operate the Browser
- LLMs in Document Intelligence: Survey, Progress, and Future Trends
- Chinese Team Trains "Spiking Large Model," Boosting Inference Speed by 100 Times
- NeurIPS'25! AutoPrune: A Plug-and-Play Adaptive Pruning Framework for Large Models
- SJTU & Stanford Propose "Long Code Compression Artifact": 5.6x Extreme Slimming Without Performance Drop
- Princeton Danqi Chen's Group's New Work: RLHF Insufficient, RLVR Bounded? RLMT Forges a Third Path
- First Code World Model Ignites AI Community, Enabling "True Reasoning" for Agents, Meta Open-Sources It
- The More You Think, The More You Err: CoT "Deep Deliberation" as a Catalyst for LLM Hallucinations!
- Boost LLM Reasoning Accuracy to 99% Without Fine-Tuning! Try DeepConf, a Lightweight Inference Framework | Latest from Meta