πŸ€– AI Research Digest – 2026-04-04

LLM

Grounded Token Initialization for New Vocabulary in LMs for Generative Recommendation

πŸ“„ Summary: This paper identifies a critical bottleneck in extending language models with new vocabulary tokens: mean initialization collapses tokens into a degenerate subspace, erasing distinctions that fine-tuning cannot fully recover. The authors use spectral and geometric analysis to diagnose this problem and propose grounded initialization strategies to better preserve inter-token distinctions for domain-specific tasks like semantic recommendation.

πŸ’‘ Key Insight: Simply averaging existing embeddings when adding new tokens to language models backfires by destroying the uniqueness you're trying to create.

πŸ”— Read Paper


Batched Contextual Reinforcement: A Task-Scaling Law for Efficient Reasoning

πŸ“„ Summary: This paper introduces a minimalist training approach where language models solve multiple problems simultaneously in a shared context window, creating an implicit token budget that forces efficient reasoning. The method discovers a novel task-scaling law showing how concurrent problem solving trades off quality and efficiency without complex training pipelines.

πŸ’‘ Key Insight: Teaching models to juggle multiple problems at once naturally teaches them to reason more efficiently without explicit penalties.

πŸ”— Read Paper


No Single Best Model for Diversity: Learning a Router for Sample Diversity

πŸ“„ Summary: The paper demonstrates that no single LLM excels at generating diverse responses across all prompts, but introduces a "diversity coverage" metric to measure answer set quality and proposes a router that learns to select the best model per-prompt for comprehensive response generation.

πŸ’‘ Key Insight: Different models have hidden strengths for different types of open-ended questionsβ€”smart routing beats picking one winner.

πŸ”— Read Paper


De Jure: Iterative LLM Self-Refinement for Structured Extraction of Regulatory Rules

πŸ“„ Summary: De Jure is a fully automated pipeline for converting dense legal documents into machine-readable regulatory rules through LLM-driven decomposition, multi-criteria evaluation, and iterative repair, requiring no human annotation or domain expertise. The system normalizes documents, structures rules semantically, judges quality across 19 dimensions, and automatically fixes low-scoring extractions within a bounded budget.

πŸ’‘ Key Insight: Legal compliance automation is possible by letting LLMs iteratively fix their own rule extraction errors without human lawyers.