"there's nothing interesting on arxiv these days!"
- the words of an uncurious mind
i have personally been blown away by the volume of interesting papers posted over the last few months, and eagerly following daily digests
here are some papers i enjoyed the most:
- Pre-training under infinite compute (September 2025, https://t.co/3Q838oO6ei)
- Fresh in memory: Training-order recency is linearly encoded in language model activations (September 2025, https://t.co/V9qCttiFPJ)
- Subliminal Learning: Language models transmit behavioral traits via hidden signals in data (July 2025, https://t.co/eJrGChfq1d)
- Memory Limitations of Prompt Tuning in Transformers (September 2025, https://t.co/AJR17dkVUx)
- Behavioral Fingerprinting of Large Language Models (September 2025, https://t.co/ZdHMlIdcYP)
- Language Self-Play For Data-Free Training (September 2025, https://t.co/9kLvY8dNbe)
- The Illusion of Diminishing Returns: Measuring Long Horizon Execution in LLMs (September 2025, https://t.co/X7bwtKE8xe)
- Do Natural Language Descriptions of Model Activations Convey Privileged Information? (September 2025, https://t.co/4qjWhFJVUG)
- Beyond the Leaderboard: Understanding Performance Disparities in Large Language Models via Model Diffing (September 2025, https://t.co/2ejyGDCSVF)
- Stochastic activations (September 2025, https://t.co/1xoXmLeIiF)
- PonderLM-2: Pretraining LLM with Latent Thoughts in Continuous Space (September 2025, https://t.co/gZW50tvCIK)
- Words That Make Language Models Perceive (October 2025, https://t.co/IDQEXdeAGv)
- Language Models Do Not Embed Numbers Continuously (October 2025, https://t.co/g8Cw3yNcoV)
- Learning Facts at Scale with Active Reading (August 2025, https://t.co/aw3fE8dKiJ)
- OverFill: Two-Stage Models for Efficient Language Model Decoding (August 2025, https://t.co/Wku5FXbGEz)
- Retrieval Capabilities of Large Language Models Scale with Pretraining FLOPs (August 2025, https://t.co/TWgqTCHjuZ)
- Reasoning-Intensive Regression (August 2025, https://t.co/2G8Lxn323A)
- Watch the Weights: Unsupervised monitoring and control of fine-tuned LLMs (August 2025, https://t.co/im0qdNorNQ)
- On the Theoretical Limitations of Embedding-Based Retrieval (August 2025, https://t.co/7haVnfNpTp)
research @cornell // language models, information theory, science of AI