Story Details

  • Atlas: Learning to Optimally Memorize the Context at Test Time

    Posted: 2025-05-31 14:13:00

    Atlas is a new approach to in-context learning that aims to optimize the selection and ordering of examples within the prompt at test time, rather than relying on heuristics or random sampling. It learns a "memorization mechanism" during training that identifies the most informative examples for a given test instance. This mechanism is implemented as a differentiable selection and ordering process, allowing it to be trained end-to-end alongside the base model. By learning which examples to include and how to arrange them, Atlas improves the effectiveness of in-context learning, achieving state-of-the-art performance on various tasks including question answering and natural language inference. This approach offers a more principled and adaptable way to leverage context within large language models compared to traditional prompt engineering.

    Summary of Comments ( 0 )
    https://news.ycombinator.com/item?id=44144407

    Hacker News users discussed the practicality and novelty of the "Atlas" model for in-context learning. Some questioned the real-world usefulness of a method that requires significant computation at test time, especially compared to simply fine-tuning a smaller model. Others highlighted the potential benefits for situations where retraining is impossible or undesirable, like personalized federated learning. The comparison to kernel methods and the potential for optimization using techniques like locality sensitive hashing were also explored. Several commenters pointed out the connection to "test-time training," a previously explored area of research, questioning the true innovation of Atlas. Finally, some found the experimental setup and evaluation unconvincing, calling for comparisons against more sophisticated baselines.