A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).
-
Updated
Jan 1, 2025
A curated reading list of research in Adaptive Computation, Inference-Time Computation & Mixture of Experts (MoE).
[Preprint] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models
Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto
Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount of time on any token
(NeurIPS-2019 MicroNet Challenge - 3rd Winner) Open source code for "SIPA: A simple framework for efficient networks"
The ARL Hierarchical MultiScale Framework (ARL-HMS) is a software library for development of multiscale models on heterogeneous high-performance computing systems.
Model implementation for "Adaptive computation as a new mechanism of human attention"
Add a description, image, and links to the adaptive-computation topic page so that developers can more easily learn about it.
To associate your repository with the adaptive-computation topic, visit your repo's landing page and select "manage topics."