You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<br>Source: *Faithful Reasoning Using Large Language Models* by Antonia Creswell et al. (2022)](https://arxiv.org/abs/2208.14271)
@@ -432,7 +432,7 @@ The method is complicated, and works as follows:
432
432
433
433
- First, build a maieutic tree, where each node is a statement that could be true or false:
434
434
- Start with a multiple-choice question or true/false statement (e.g. `War cannot have a tie`)
435
-
- For each possible answer to the question, use the model to generate a correponding explanation (with a prompt like `War cannot have a tie? True, because`)
435
+
- For each possible answer to the question, use the model to generate a corresponding explanation (with a prompt like `War cannot have a tie? True, because`)
436
436
- Then, prompt the model with the question and the generated explanation, and ask it to produce the answer. If reversing the explanation (with a prefix like `It is wrong to say that {explanation}`) reverses the answer, then the explanation is considered 'logically integral.'
437
437
- If an explanation is not logically integral, then repeat the above process recursively, with each explanation turned into a True or False question, and generate more explanations for each new question.
438
438
- After all of the recursive explaining is done, you end up with a tree of explanations, where each leaf on the tree has the property that reversing the explanation reverses the model's answer.
@@ -505,7 +505,7 @@ In 2021, OpenAI researchers applied this technique to grade school math problems
505
505
506
506
#### Results
507
507
508
-
With a 175B GPT-3 model and 8,000 training examples, this technique substantially lifted gradeschool math accuracy from ~33% to ~55%.
508
+
With a 175B GPT-3 model and 8,000 training examples, this technique substantially lifted grade school math accuracy from ~33% to ~55%.
509
509
510
510
[
511
511
<br>Source: *Training Verifiers to Solve Math Word Problems* by Karl Cobbe et al. (2021)](https://arxiv.org/abs/2110.14168)
@@ -571,4 +571,4 @@ In the future, expect better models and better techniques to be published. Even
571
571
| On long reasoning problems, you can improve step-by-step reasoning by splitting the problem into pieces to solve incrementally |[Least-to-most Prompting Enables Complex Reasoning in Large Language Models](https://arxiv.org/abs/2205.10625)| 2022 May |
572
572
| You can have the model analyze both good and bogus explanations to figure out which set of explanations are most consistent |[Maieutic Prompting: Logically Consistent Reasoning with Recursive Explanations](https://arxiv.org/abs/2205.11822)| 2022 May |
573
573
| You can think about these techniques in terms of probabilistic programming, where systems comprise unreliable components |[Language Model Cascades](https://arxiv.org/abs/2207.10342)| 2022 Jul |
574
-
| You can eliminate hallucination with sentence label manipulation, and you can reduce wrong answers with a 'halter' prompt |[Faithful Reasoning Using Large Language Models](https://arxiv.org/abs/2208.14271)| 2022 Aug |
574
+
| You can eliminate hallucination with sentence label manipulation, and you can reduce wrong answers with a 'halter' prompt |[Faithful Reasoning Using Large Language Models](https://arxiv.org/abs/2208.14271)| 2022 Aug |
0 commit comments