Discussion about this post

User's avatar
Meng Li's avatar

LLMs may not solve problems correctly on the first try, especially for tasks they haven't been trained on.

To make the most of these models, they must be able to do two things: 1. Identify where their reasoning went wrong; 2. Backtrack to find another solution.

This has led to a surge in methods related to LLM self-correction in the industry, which involves using LLMs to identify issues in their outputs and then generating improved results based on feedback.

Self-correction is often considered a single process, but it can be broken down into two parts: error detection and output correction.

Expand full comment
1 more comment...

No posts