Kecheng Greetings!
As the title suggests, can RLHF be included in the prompt flow to improve the quality of outputs? If not, what other ways can I improve on the quality of outputs? (other than modifying the prompts)
I haven't found any information that specifically addresses your question about including Reinforcement Learning with Human Feedback (RLHF) into Azure prompt flow. However, please check below information and let me know if that helps.
One way to improve the quality of outputs is to use Few-shot, one-shot and zero-shot approaches for in-context learning. These approaches vary based on the amount of task-specific data that is given to the model. Few-shot approach is where a user includes several examples in the call prompt that demonstrate the expected answer format and content. One-shot approach is where a user includes a single example in the call prompt that demonstrates the expected answer format and content. Zero-shot approach is where a user provides no examples in the call prompt, and the model generates a response based on its training data.
Please see Prompt engineering techniques, Get started with prompt flow and Tune prompts using variants for more details.