Leveraging Retrieval-Augmented Prompting for Enhanced Comment Feedback Prediction with Large Language Models

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

User comment feedback prediction is an important problem in online platforms, which have impact on the automatic moderation, quality evaluation, and user experience enhancement etc. Despite strong general reasoning capabilities of large language models (LLMs) such as GPT-5, Claude, Gemini and Qwen3-7B, their performance on domain restriction tasks can vary significantly and is commonly very sensitive to context provided and external knowledge. This paper presents an innovative Retrieval-Augmented Prompting (RAP) framework for comment feedback prediction. Our approach incrementally improves the quality of LLM prompts by retrieving k of the semantically most similar historical comment-feedback pairs from an external dataset as in-context few-shot examples. We experiment on the specialized dataset of 50000 comment ratings, obtained from various online scenarios. Experimental results show that our GPT-5 + RAP model outperforms state-of-the-art LLMs such as Qwen3-7B, Claude, Gemini and GPT-5 baseline on accuracy, Macro-F1 and explanation consistency and a strong prompt engineered baseline, GPT-5 + PE.

Article activity feed