The Grand AI Handbook
The Grand AI Handbook
March 18-24, 2025
Models Reasoning Reinforcement Learning Agents

Week 12: Recent Advances in LLM Research

This collection highlights cutting-edge research in LLM architecture, reinforcement learning approaches, and scaling dynamics. Featured papers explore innovations in attention mechanisms, hierarchical reward models, and specialized memory systems for advanced reasoning capabilities.

Research Highlights

A Review of DeepSeek Models

Anonymous Paper Link

An in-depth review of techniques behind DeepSeek's open-source LLMs—DeepSeek-V3 and DeepSeek-R1—which achieve state-of-the-art performance with lower resource requirements.

  • Multi-Head Latent Attention (MLA) compresses keys and values into latent vectors, reducing memory consumption
  • Advanced Mixture of Experts (MoE) with fine-grained segmentation and dedicated shared experts
  • Group Relative Policy Optimization (GRPO) eliminates value function approximation from PPO

"DeepSeek's approach demonstrates how algorithm-hardware co-design can maximize computational efficiency while achieving cutting-edge performance."

Hierarchical Multi-Step Reward Models for Enhanced Reasoning

Anonymous Paper Link

This work proposes a Hierarchical Reward Model (HRM) that addresses reward hacking and error propagation issues in fine-grained LLM reasoning.

  • Assesses multiple consecutive reasoning steps rather than individual steps
  • Introduces Hierarchical Node Compression (HNC) to augment MCTS-based data annotation
  • Outperforms standard reward models on PRM800K and cross-domain tasks

"HRM's multi-step feedback framework mitigates reward hacking behaviors by penalizing incomplete or incoherent reasoning across sequences."

DAPO: Open-Source LLM Reinforcement Learning System

Anonymous Paper Link

DAPO presents a fully open-source, large-scale RL system that boosts chain-of-thought reasoning capabilities in LLMs through several innovative techniques.

  • "Clip-Higher" approach in PPO-style training prevents entropy collapse
  • Filters training samples to focus on those with useful gradient signals
  • Achieves SOTA math performance on AIME 2024 with 50% accuracy

"DAPO outperforms DeepSeek's R1 with less training time, showcasing open-source reproducibility at scale for advanced reasoning capabilities."

Compute Optimal Scaling of Skills

University of Wisconsin and Meta AI Paper Link

This research investigates how different skills exhibit contrasting optimal scaling behaviors in LLMs, revealing distinct preferences for model size versus data volume.

  • Knowledge tasks prefer bigger models (capacity-hungry)
  • Code tasks prefer more data tokens (data-hungry)
  • Validation set choice can misalign compute-optimal model sizes by 30–50%

"Model developers must design validation sets that represent the target skill mix to optimize for the correct parameter-to-data ratio."

Thinking Machines: Survey of Reasoning Techniques

Anonymous Paper Link

This survey provides a comprehensive overview and comparison of existing reasoning techniques in language models.

  • Systematically categorizes reasoning-imbued language models
  • Compares different approaches to enhancing reasoning capabilities
  • Identifies trends and future directions in reasoning research

"The survey establishes a framework for understanding the rapidly evolving landscape of reasoning techniques in modern LLMs."

A Survey on Efficient Reasoning

Anonymous Paper Link

This survey investigates techniques to address the "overthinking phenomenon" in Large Reasoning Models (LRMs).

  • Categorizes methods into model-based, output-based, and prompt-based optimizations
  • Explores the balance between reasoning capability and computational efficiency
  • Examines approaches used in models like OpenAI o1 and DeepSeek-R1

"The survey highlights ongoing efforts to make advanced reasoning more computationally efficient without sacrificing performance."

A-MEM: Agentic Memory for LLM Agents

Rutgers University and Ant Group Paper Link

A-MEM introduces a Zettelkasten-inspired memory system for LLM agents that enables dynamic, evolving knowledge representation for complex tasks.

  • Creates comprehensive memory notes with textual attributes and embeddings
  • Automatically updates older memories when new related information arrives
  • Outperforms static-memory methods like MemGPT while reducing token usage

"A-MEM's continuous memory evolution enables a more coherent, ever-improving knowledge network capable of capturing deeper connections over time."

DeepMesh: Artist-like 3D Mesh Generation

Tsinghua University, Nanyang Technological University, ShengShu Paper Link

DeepMesh is a transformer-based system that generates high-quality 3D meshes with artist-like topology, balancing efficiency and aesthetic quality.

  • Compresses mesh sequences by ~72% while preserving geometric detail
  • Uses Direct Preference Optimization (DPO) to align with human aesthetic preferences
  • Handles large meshes and supports both point cloud and image-based conditioning

"DeepMesh predicts structured triangle layouts that are both aesthetic and easy to edit, outperforming baselines in geometric accuracy and user ratings."

Deep Learning is Not So Mysterious or Different

Andrew Gordon Wilson (New York University) Paper Link

This perspective paper argues that phenomena in deep learning such as benign overfitting and double descent are neither mysterious nor exclusive to neural networks.

  • Demonstrates benign overfitting with simple linear models and polynomials
  • Advocates for soft inductive biases instead of traditional hard constraints
  • Shows how established frameworks already explain supposedly puzzling behaviors

"The paper critiques neural network exceptionalism, urging closer collaboration between communities to build on established generalization theories rather than reinventing them."

GNNs as Predictors of Agentic Workflow Performances

Anonymous Paper Link

This work introduces FLORA-Bench, a large-scale benchmark to evaluate GNN-based predictors for automating and optimizing agentic workflows.

  • Graph Neural Networks efficiently predict success of multi-agent LLM workflows
  • Significantly reduces costly repeated model calls
  • Provides a framework for optimizing complex agent systems

"FLORA-Bench demonstrates how graph-based approaches can make agentic systems more efficient by predicting which workflows are likely to succeed."

Emerging Trends

Industry Implications

This research collection offers significant implications for AI applications:

Resource-Efficient Deployment

Techniques like Multi-Head Latent Attention and optimized RL approaches enable high-performance models with lower infrastructure requirements.

Specialized Training Strategies

Understanding that different skills require different scaling approaches allows for more targeted allocation of computational resources.

Enhanced Reasoning Reliability

Hierarchical reward models and better reasoning frameworks promise more reliable AI systems for critical decision-making contexts.

Agent Infrastructure Optimization

Tools like FLORA-Bench and A-MEM support more efficient development of complex agent systems with better memory and workflow prediction.