The Silent Onset of an AI-Scored Society — How Conclusions Without Process Quietly Reallocate Social Visibility
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
The purpose of this paper is to formalize and test a hypothesis about how generative and recommendation AI systems are silently reallocating social visibility through automated scoring of discourse. We argue that such systems are reshaping human trust structures without visible interfaces, creating a second wave of algorithmic gatekeeping. Our contributions are threefold: (1) we introduce a two-wave model that distinguishes the visible displacement of human labor by AI tools from the invisible redistribution of attention via algorithmic recommendation, highlighting the societal importance of this shift; (2) we define three measurable indicators—Structure Index (SI), Reusability Index (RI), and Closure Rate (CR)—to operationalize the process quality of discourse and enable quantitative study; and (3) we derive falsifiable predictions suggesting that high–SI/RI/CR discourse is more likely to remain visible, while conclusion-only content is systematically down-scored. We further argue that nonproductive escalation signals (anger, provocation, spike-type virality) degrade long-term visibility. This work bridges algorithmic fairness research with the emerging need to understand AI-mediated social gatekeeping, and it provides a foundation for future empirical validation and platform design intervention.