-
iTTT (WIP)
Implicit Test-Time Training for infinite-context sequence modelling and continual learning. Currently a work in progress.
-
ZEBRA (WIP)
Train parallel and test serial for scalable latent reasoning. Currently a work in progress.
-
CWIC
Compute Where It Counts. A new state-of-the-art method for creating sparse transformers that automatically decide when to use more or less compute.
-
pBit
Enabling both sparsity and low-bit quantization in neural networks using stochastic weights and the local reparameterization trick.
-
MonArc
Efficiently pretraining energy-based language models that operate at the token level.
-
NoiseSearch
Improving diffusion model sample quality by searching for better random noise using metaheuristics (early work on test-time scaling).