For setting parameters, OpenAI recommends using ada, as it is the fastest engine and capable of creating good predictions in a classification task after fine-tuning. At least 100 examples per class are needed for fine-tuning, and the performance tends to increase linearly as the number of examples doubles.
To get confidence, the log probability of the first generated completion token can be used. The higher the log probability for the "yes" token, the more confident the prediction is that the output is supported.
To determine a log probability threshold above which the ad is likely to be supported more than 98% of the time, follow these steps:
- Use the discriminator to predict the probability of "yes" on a held-out dataset.
- Convert the log probability measure into percentiles.
- For each percentile compute a precision, which is the share of actual truthful ads found above that threshold.
- Find a percentile at which the precision is just above 98%. The log probability threshold needed to obtain a precision of at least 98% is then the log probability at this percentile on the held-out dataset.
By generating several samples and then picking the one with the highest log probability, you can increase the probability that the selected ad is indeed truthful. For example, by generating 3 samples you can achieve a sufficient truthfulness threshold of 98.3%. This can be increased to 99.6% by generating 10 samples, and 99.8% by generating 17 samples. There are diminishing returns as the number of samples significantly increases.