Discussion about this post

User's avatar
Derek Lomas's avatar

I'm really interested in the user interface design of these human feedback systems. Reading 64 outputs to choose the best? How do people do that?

Expand full comment
Ilse Langnar's Notebook's avatar

I wonder if the model could suffer from initial errors in the training data and then the feedback-loop would make it become "super-wrong". I wonder how things such as Dataset poising and mallicious attacks might be played out, either in the data or in the RL by people.

Expand full comment
4 more comments...

No posts