Prompt
Last updated
Was this helpful?
Last updated
Was this helpful?
Brex on , but goes through the history of language models which is amazing
, a good summary of all the techniques
- this is a very thorough review of the topic
- optimizing over a set of candidate that were proposed by an LLM in order to maximize a score function. contributes to improvement of responses.
- given the variability of the quality of results, how do we pick the best prompts automatically? using GPT3 and back translation to choose the lowest perplexity prompts that give the most gain in performance.
- it becomes more competitive at scale.
,
(amazing) ,
Prompting Elicits Reasoning in Large Language Models - COT is a series of intermediate reasoning steps that significantly improves the ability of large language models to perform complex reasoning, by Jason Wei Xuezhi Wang Dale Schuurmans Maarten Bosma Brian Ichter Fei Xia Ed H. Chi Quoc V. Le Denny Zhou Google Research, Brain Team.
- "samples a diverse set of reasoning paths instead of only taking the greedy one, and then selects the most consistent answer by marginalizing out the sampled reasoning paths"
- Step-Back Prompting (STP) is prompt approach in which we teach the model to answer a global questions, i.e., the original question is transformed into a stepback question, and the answer to the stepback question is used to formulate the final response.