Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning

Research output: Chapter in book/report/conference proceedingConference abstractResearchpeer review

View graph of relations

Details

Original languageEnglish
Title of host publication2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025)
Publication statusAccepted/In press - 15 Feb 2025

Abstract

While increasingly large models have revolutionized much of the machine learning landscape, training even mid-sized networks for Reinforcement Learning (RL) is still proving to be a struggle. This, however, severely limits the complexity of policies we are able to learn. To enable increased network capacity while maintaining network trainability, we propose GrowNN, a simple yet effective method that utilizes progressive network growth during training. We start training a small network to learn an initial policy. Then we add layers without changing the encoded function. Subsequent updates can utilize the added layers to learn a more expressive policy, adding capacity as the policy’s complexity increases. GrowNN can be seamlessly integrated into most existing RL agents. Our experiments on MiniHack and Mujoco show improved agent performance, with incrementally GrowNN deeper networks outperforming their respective static counterparts of the same size by up to 48% on MiniHack Room and 72% on Ant.

Cite this

Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning. / Fehring, Lukas; Eimer, Theresa; Lindauer, Marius.
2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025). 2025.

Research output: Chapter in book/report/conference proceedingConference abstractResearchpeer review

Fehring, L, Eimer, T & Lindauer, M 2025, Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning. in 2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025).
Fehring, L., Eimer, T., & Lindauer, M. (Accepted/in press). Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning. In 2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025)
Fehring L, Eimer T, Lindauer M. Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning. In 2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025). 2025
Fehring, Lukas ; Eimer, Theresa ; Lindauer, Marius. / Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning. 2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025). 2025.
Download
@inbook{814dda21f5164ae79e66e0d5975cb457,
title = "Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning",
abstract = "While increasingly large models have revolutionized much of the machine learning landscape, training even mid-sized networks for Reinforcement Learning (RL) is still proving to be a struggle. This, however, severely limits the complexity of policies we are able to learn. To enable increased network capacity while maintaining network trainability, we propose GrowNN, a simple yet effective method that utilizes progressive network growth during training. We start training a small network to learn an initial policy. Then we add layers without changing the encoded function. Subsequent updates can utilize the added layers to learn a more expressive policy, adding capacity as the policy{\textquoteright}s complexity increases. GrowNN can be seamlessly integrated into most existing RL agents. Our experiments on MiniHack and Mujoco show improved agent performance, with incrementally GrowNN deeper networks outperforming their respective static counterparts of the same size by up to 48% on MiniHack Room and 72% on Ant.",
author = "Lukas Fehring and Theresa Eimer and Marius Lindauer",
year = "2025",
month = feb,
day = "15",
language = "English",
booktitle = "2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025)",

}

Download

TY - CHAP

T1 - Growing with Experience: Growing Neural Networks in Deep Reinforcement Learning

AU - Fehring, Lukas

AU - Eimer, Theresa

AU - Lindauer, Marius

PY - 2025/2/15

Y1 - 2025/2/15

N2 - While increasingly large models have revolutionized much of the machine learning landscape, training even mid-sized networks for Reinforcement Learning (RL) is still proving to be a struggle. This, however, severely limits the complexity of policies we are able to learn. To enable increased network capacity while maintaining network trainability, we propose GrowNN, a simple yet effective method that utilizes progressive network growth during training. We start training a small network to learn an initial policy. Then we add layers without changing the encoded function. Subsequent updates can utilize the added layers to learn a more expressive policy, adding capacity as the policy’s complexity increases. GrowNN can be seamlessly integrated into most existing RL agents. Our experiments on MiniHack and Mujoco show improved agent performance, with incrementally GrowNN deeper networks outperforming their respective static counterparts of the same size by up to 48% on MiniHack Room and 72% on Ant.

AB - While increasingly large models have revolutionized much of the machine learning landscape, training even mid-sized networks for Reinforcement Learning (RL) is still proving to be a struggle. This, however, severely limits the complexity of policies we are able to learn. To enable increased network capacity while maintaining network trainability, we propose GrowNN, a simple yet effective method that utilizes progressive network growth during training. We start training a small network to learn an initial policy. Then we add layers without changing the encoded function. Subsequent updates can utilize the added layers to learn a more expressive policy, adding capacity as the policy’s complexity increases. GrowNN can be seamlessly integrated into most existing RL agents. Our experiments on MiniHack and Mujoco show improved agent performance, with incrementally GrowNN deeper networks outperforming their respective static counterparts of the same size by up to 48% on MiniHack Room and 72% on Ant.

M3 - Conference abstract

BT - 2025 Multi-disciplinary Conference on Reinforcement Learning and Decision Making (RLDM 2025)

ER -

By the same author(s)