Recent Preprints

You can also find my articles on my Google Scholar profile.

A self-improving coding agent

Published in ICLR Workshop on Scaling Self-Improving Foundation Models without Human Supervision (Oral), 2025

This paper presents a coding agent that can improve itself without human supervision.

Recommended citation: Robeyns M, Szummer M, and Aitchison L. (2025). "A self-improving coding agent." ICLR Workshop on Scaling Self-Improving Foundation Models without Human Supervision (Oral).

Questionable practices in machine learning

Published in arXiv, 2024

This paper examines questionable research practices in the field of machine learning.

Recommended citation: Leech G, Vazquez JJ, Kupper N, Yagudin M, and Aitchison L. (2024). "Questionable practices in machine learning." arXiv:2407.12220.
Download Paper

Batch size invariant Adam

Published in arXiv, 2024

This paper introduces a variant of Adam that is invariant to batch size.

Recommended citation: Wang X and Aitchison L. (2024). "Batch size invariant Adam." arXiv:2402.18824.
Download Paper

Function-space learning rates

Published in arXiv, 2024

This paper explores learning rates in function space.

Recommended citation: Milsom E, Anson B, and Aitchison L. (2024). "Function-space learning rates." arXiv:2502.17405.
Download Paper

Bayesian reward models for LLM alignment

Published in Workshop on Secure and Trustworthy Large Language Models, 2024

This paper explores Bayesian reward models for aligning large language models.

Recommended citation: Yang AX, Robeyns M, Coste T, Wang J, Bou-Ammar H, and Aitchison L. (2024). "Bayesian reward models for LLM alignment." Workshop on Secure and Trustworthy Large Language Models.

LoRA ensembles for large language model fine-tuning

Published in NeurIPS Workshop: Efficient Natural Language and Speech Processing, 2023

This paper investigates the use of LoRA ensembles for fine-tuning large language models.

Recommended citation: Wang X, Aitchison L, and Rudolph M. (2023). "LoRA ensembles for large language model fine-tuning." NeurIPS Workshop: Efficient Natural Language and Speech Processing.