@ericjang11
Eric Jang
1 year
Instead of finding the perfect prompt for an LLM (let's think step by step), you can ask LLMs to critique their outputs and immediately fix their own mistakes. Here's a fun example:
61
224
2K

Replies

@ericjang11
Eric Jang
1 year
I saw @awjuliani 's tweet that LLMs cannot generate a non-rhyming poem. Indeed, GPT-4 does not do it even if I ask it to think carefully
Tweet media one
@awjuliani
Arthur Juliani
1 year
Still can't not rhyme.
Tweet media one
3
1
24
3
6
77
@ericjang11
Eric Jang
1 year
@awjuliani Now you ask GPT-4 if it met the assignment, at which point it apologizes and generates a valid non-rhyming poem! full marks
Tweet media one
19
12
326
@ericjang11
Eric Jang
1 year
@awjuliani h/t @avisingh599 for pointing me out the Reflexion paper, which gets at this idea. It's absolutely wild to me that LLMs are general enough that they can critique their own outputs in a sensible way
6
14
307
@ericjang11
Eric Jang
1 year
@awjuliani @avisingh599 Interestingly enough, GPT-3.5 is incapable of such self-critique, at least for this assignment. It seems to be an emergent capability only present in GPT-4
Tweet media one
15
14
225
@ericjang11
Eric Jang
1 year
@awjuliani @avisingh599 The implications are substantial. Instead of clever "prefix prompt engineering", we can now consider a "postfix prompt engineering", which encourages LLMs to find corrections and inconsistencies within prior generated solutions.
4
25
312
@ericjang11
Eric Jang
1 year
@awjuliani @avisingh599 After any generated output, just append "did the generated output do what the user asked?" and the LLM becomes a "minimal policy improvement operator" for itself
4
20
177
@ericjang11
Eric Jang
1 year
@awjuliani @avisingh599 Maybe it is possible to apply a critique to the critique recursively, i.e. append "is the critique logically consistent with the original request?" GPT-3.5 seems to be rather biased towards self-congratulatory optimism
Tweet media one
8
5
70
@ericjang11
Eric Jang
1 year
Wrote a quick blog post about it here. Would be curious to see what examples it can verify & correct, and what examples it can only verify, and what tasks it fails to verify
5
12
110
@ericjang11
Eric Jang
1 year
Also this self-critique trick seems to be able to handle @ylecun 's gear puzzle!
@ankesh_anand
Ankesh Anand
1 year
@ericjang11 @stanislavfort @ylecun works if you ask it to critique itself :)
Tweet media one
4
0
26
5
2
56
@ericjang11
Eric Jang
1 year
@character_ai 's c.ai @1 .2 does a pretty good job without further prefix or postfix prompting
Tweet media one
3
0
16
@C_H_Wood
Chris Wood
1 year
@ericjang11 You should try asking it what prompt it would create to get itself (or I’ve found GPT3 believes that only GPT2 exists so I use that when I do this) to do what you want. Then you show it the results and share your critique and ask it how to change the prompt to improve the output
1
0
3
@tljstewart
Timothy L.J. Stewart
1 year
@ericjang11 Reprompt FTW
0
1
3
@steventlive
Steven Talafous
1 year
@ericjang11 Yup. Extremely apparent and useful when asking to generate/enhance code.
0
0
2
@Lucca_Scarpa_
Lucca Scarpa
1 year
1
0
0
@enthrighte
enthrighte
1 year
@ericjang11 @lndian_Bronson ??? Did everyone not read the poem? It still rhymes ?
0
0
1
@mindaria
Spyros Kalaitzis 🚢
1 year
1
0
0
@_jumski
Jumski Jumster
1 year
1
0
0
@hellolufeng
薛定鳄的校长
1 year
1
0
0
@ymyumo
语默
1 year
1
0
0
@sridatta
Midnight Maniac Sri
1 year
@ericjang11 @ShreyaR Is this technique something that could be used in RAIL?
1
0
0