Hace 1 año | Por alafia a zeta-alpha.com
Publicado hace 1 año por alafia a zeta-alpha.com

OpenAI ha sacudido el mundo de la Inteligencia Artificial con su ChatGPT en Noviembre de 2022. Su artículo científico de Marzo de 2022, “Training language models to follow instructions with human feedback” ha sido uno de los más citados en 2022. El avance de este campo es tan rápido que es muy difícil estar al corriente de los últimos descubrimientos. Esta entrada recoge los 100 artículos científicos más importantes del año pasado ordenados por el número de citas. El más citado es de DeepMind (Google), y su IA Alphafold. [Ver #1]

Comentarios

alafia

El ya mítico artículo de OpenAI:
Training language models to follow instructions with human feedback
Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, Ryan Lowe

Making language models bigger does not inherently make them better at following a user's intent. For example, large language models can generate outputs that are untruthful, toxic, or simply not helpful to the user. In other words, these models are not aligned with their users. In this paper, we show an avenue for aligning language models with user intent on a wide range of tasks by fine-tuning with human feedback. Starting with a set of labeler-written prompts and prompts submitted through the OpenAI API, we collect a dataset of labeler demonstrations of the desired model behavior, which we use to fine-tune GPT-3 using supervised learning. We then collect a dataset of rankings of model outputs, which we use to further fine-tune this supervised model using reinforcement learning from human feedback. We call the resulting models InstructGPT. In human evaluations on our prompt distribution, outputs from the 1.3B parameter InstructGPT model are preferred to outputs from the 175B GPT-3, despite having 100x fewer parameters. Moreover, InstructGPT models show improvements in truthfulness and reductions in toxic output generation while having minimal performance regressions on public NLP datasets. Even though InstructGPT still makes simple mistakes, our results show that fine-tuning with human feedback is a promising direction for aligning language models with human intent.

https://arxiv.org/abs/2203.02155

El más citado de 2022:

AlphaFold Protein Structure Database: massively expanding the structural coverage of protein-sequence space with high-accuracy models
https://academic.oup.com/nar/article/50/D1/D439/6430488

Para todo lo demás cof*SciHub*cof.