Using Synthetic Persona Generation to Improve Prompt Responses
AutoPrompt uses GPT to generate synthetic personas and their relevant documents and asking the personas to compare their prompt results with one that had context from their documents. Here's a quick result from this approach with 79 personas.
From this, we had to investigate the possibility of over-fitting to under-represented personas and ran two experiments to test that overfitting hypothesis over this prototype:
A: For classes of prompts, the baseline GPT response was 17% less satisfactory for underrepresented personas. B: Using the content of the persona’s documents is 11% more satisfactory than explicit demographic context.
These suggest that adding context from documents may overcome gaps in the underlying baseline of GPT’s recommendation responses (A) but may not be explicitly overfitting to the demographics (B). Some suggested next steps include:
- Regularizing with relevant global context.
- Incorporating real data.
- Testing with a larger sample size.
There's definitely a lot of interesting directions going forward here, but these types of synthetic setups enable different types of experiementations than what was historically possible.