Chain-of-Thought
Last updated
Was this helpful?
Last updated
Was this helpful?
Due to scaling law of AI Model, LLMs are getting bigger and bigger. However, for reasoning tasks, performance improvement was poor.
Researchers came up with an idea of instead of changing the model checkpoints, they tried to make the model mimic human's thinking process using prompt engineering.
Chain-of-Thought technique is making a model use a reasoning-chain like human.
Instead of making model just output the answer, we can make the model output all the reasoning path approaching to the answer. Since LLMs(Transformer) are auto-regressive models, it reference the previous reasoning path to approach to the answer.
[1]
[2]