Settings

Theme

New research shows RL may not help a model learn new basic skills

arxiv.org

1 points by binsquare 19 days ago · 1 comment

Reader

binsquareOP 19 days ago

Reinforcement learning dominated the recent neurips papers. But here's one that stood out to me about how exactly pre training can affect post training.

This means if the core data (ex. additions, subtractions, etc) were not there in the pre training stage, RL on complex math problems would not lead to the model developing improvements in the core areas.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection