选择性说服:生成式人工智能在争议性议题中的结构性立场偏见Selective Persuasion: Structural Stance Bias of Generative AI in Controversial Issues
黎樟浩,塔娜
摘要(Abstract):
生成式人工智能日益成为公众获取信息的重要工具,其基于用户特征进行内容定制的个性化说服能力尤为突出,同时也引发了对算法偏见的广泛关注。本文以国内三个主流生成式人工智能平台为研究对象,围绕公共健康、公共政策与社会伦理等争议性议题,构建涵盖性别、年龄等六个人口属性维度共96种用户画像,并据此设计提示词以评估生成内容的说服力差异。研究发现:在相同用户画像条件下,针对不同立场生成的文本存在显著的说服力差异;不同的用户画像显著影响生成文本的说服力;且人口属性对不同立场倾向的内容说服力具有调节作用。据此,本文提出“选择性说服”,揭示了生成式人工智能在争议性议题中的结构性立场偏见。这一机制体现为,生成式人工智能为不同群体或不同立场生成说服力具有显著差异的内容,可能潜在地参与公众态度形塑与社会共识构建的过程,并加剧数字不平等与群体极化等社会风险。
关键词(KeyWords): 生成式人工智能;选择性说服;个性化说服;算法审计;数字不平等
基金项目(Foundation): 新一代人工智能国家科技重大专项“人工智能社会实验伦理、评估与标准化研究”(项目编号:2023ZD0121600)课题四“智能社会治理仿真演化模型研究”(项目编号:2023ZD0121604);; 中央高校基本科研业务费专项资金资助的阶段性成果
作者(Author): 黎樟浩,塔娜
参考文献(References):
- 沈伟伟(2019):算法透明原则的迷思——算法规制理论的批判,《环球法律评论》,第41卷第6期,20-39页。
- 师文、陈昌凤(2023):平台算法的“主流化”偏向与“个性化”特质研究——基于计算实验的算法审计,《新闻记者》,第11期,3-14页。
- 塔娜、林聪(2023):点击搜索之前:针对搜索引擎自动补全算法偏见的实证研究,《国际新闻界》,第45卷第8期,132-154页。
- 新京报(2025年7月10日):中国AI大模型测评报告出炉,8款大模型首测伦理判断能力,获取自 https://www.bjnews.com.cn/detail/1752140099129381.html
- 徐明华、魏子瑶(2023):算法伦理的治理新范式:算法审计的兴起、发展与未来,《当代传播》,第1期,80-86页。
- Bai,Y.T.,Jones,A.,Ndousse,K.,Askell,A.,Chen,A.N.,DasSarma,N.,Drain,D.,Fort,S.,Ganguli,D.,Henighan,T.,Joseph,N.,Kadavath,S.,Kernion,J.,Conerly,T.,El-Showk,S.,Elhage,N.,Hatfield-Dodds,Z.,Hernandez,D.,Hume,T.,Johnston,S.,Kravec,S.,Lovitt,L.,Nanda,N.,Olsson,C.,Amodei,D.,Brown,T.,Clark,J.,McCandlish,S.,Olah,C.,Mann,B.& Kaplan,J.(2022).Training a helpful and harmless assistant with reinforcement learning from human feedback.arXiv preprint arXiv:2204.05862.doi:10.48550/arXiv.2204.05862.
- Bender,E.M.,Gebru,T.,McMillan-Major,A.& Shmitchell,S.(2021).On the dangers of stochastic parrots:Can language models be too big?.In Proceedings of the 2021 ACM conference on fairness,accountability,and transparency (pp.610-623).ACM.doi:10.1145/3442188.3445922.
- Bertrand,M.& Mullainathan,S.(2004).Are Emily and Greg more employable than Lakisha and Jamal?A field experiment on labor market discrimination.American Economic Review,94(4),991-1013.doi:10.1257/0002828042002561.
- Bigman,Y.E.,Wilson,D.,Arnestad,M.N.,Waytz,A.& Gray,K.(2023).Algorithmic discrimination causes less moral outrage than human discrimination.Journal of Experimental Psychology:General,152(1),4-27.doi:10.1037/xge0001250.
- Buolamwini,J.& Gebru,T.(2018).Gender shades:Intersectional accuracy disparities in commercial gender classification.In Proceedings of the 1st Conference on Fairness,Accountability and Transparency (pp.77-91).New York,NY,USA:PMLR.
- Cheryan,S.& Markus,H.R.(2020).Masculine defaults:Identifying and mitigating hidden cultural biases.Psychological Review,127(6),1022-1052.doi:10.1037/rev0000209.
- Cialdini,R.B.(2001).The science of persuasion.Scientific American,284(2),76-81.doi:10.1038/scientificamerican0201-76.
- Coppock,A.,Ekins,E.& Kirby,D.(2018).The long-lasting effects of newspaper op-eds on public opinion.Quarterly Journal of Political Science,13(1),59-87.doi:10.1561/100.00016112.
- Coppock,A.,Hill,S.J.& Vavreck,L.(2020).The small effects of political advertising are small regardless of context,message,sender,or receiver:Evidence from 59 real-time randomized experiments.Science Advances,6(36),eabc4046.doi:10.1126/sciadv.abc4046.
- Dai,S.H.,Xu,C.,Xu,S.C.,Pang,L.,Dong,Z.H.& Xu,J.(2024).Bias and unfairness in information retrieval systems:New challenges in the LLM era.In Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining (pp.6437-6447).Barcelona,Spain:ACM.doi:10.1145/3637528.3671458.
- Dehnert,M.& Mongeau,P.A.(2022).Persuasion in the age of Artificial Intelligence (AI):Theories and complications of AI-based persuasion.Human Communication Research,48(3),386-403.doi:10.1093/hcr/hqac006.
- Farrell,H.,Gopnik,A.,Shalizi,C.& Evans,J.(2025).Large AI models are cultural and social technologies.Science,387(6739),1153-1156.doi:10.1126/science.adt9819.
- Gallegos,I.O.,Rossi,R.A.,Barrow,J.,Tanjim,M.M.,Kim,S.,Dernoncourt,F.,Yu,T.,Zhang,R.Y.& Ahmed,N.K.(2024).Bias and fairness in large language models:A survey.Computational Linguistics,50(3),1097-1179.doi:10.1162/coli_a_00524.
- Goldstein,J.A.,Chao,J.,Grossman,S.,Stamos,A.& Tomz,M.(2024).How persuasive is AI-generated propaganda?.PNAS Nexus,3(2),pgae034.doi:10.1093/pnasnexus/pgae034.
- Gorwa,R.,Binns,R.& Katzenbach,C.(2020).Algorithmic content moderation:Technical and political challenges in the automation of platform governance.Big Data & Society,7(1),1-15.doi:10.1177/2053951719897945.
- Gu,J.W.,Jiang,X.H.,Shi,Z.C.,Tan,H.X.,Zhai,X.H.,Xu,C.J.,Li,W.,Shen,Y.H.,Ma,S.J.,Liu,H.H.,Wang,S.Z.,Zhang,K.,Wang,Y.Z.,Gao,W.,Ni,L.& Guo,J.(2025).A survey on LLM-as-a-judge.arXiv preprint arXiv:2411.15594.doi:10.48550/arXiv.2411.15594.
- Hackenburg,K.& Margetts,H.(2024).Evaluating the persuasive influence of political microtargeting with large language models.Proceedings of the National Academy of Sciences of the United States of America,121(24),e2403116121.doi:10.1073/pnas.2403116121.
- Hartmann,J.,Schwenzow,J.& Witte,M.(2023).The political ideology of conversational AI:Converging evidence on ChatGPT's pro-environmental,left-libertarian orientation.arXiv preprint arXiv:2301.01768.doi:10.48550/arXiv.2301.01768.
- Hendrycks,D.,Mazeika,M.& Woodside,T.(2023).An overview of catastrophic AI risks.arXiv preprint arXiv:2306.12001.doi:10.48550/arXiv.2306.12001.
- Hirsh,J.B.,Kang,S.K.& Bodenhausen,G.V.(2012).Personalized persuasion:Tailoring persuasive appeals to recipients' personality traits.Psychological Science,23(6),578-581.doi:10.1177/0956797611436349.
- Kapoor,A.& Kumar,M.(2025).Generative AI and personalized video adver-tisements.Marketing Science,44(4),733-973.doi:10.1287/mksc.2023.0494.
- Kaptein,M.,Markopoulos,P.,de Ruyter,B.& Aarts,E.(2015).Personalizing persuasive technologies:Explicit and implicit personalization using persuasion profiles.International Journal of Human-Computer Studies,77,38-51.doi:10.1016/j.ijhcs.2015.01.004.
- Li,M.,Chen,J.H.,Chen,L.C.& Zhou,T.Y.(2024).Can LLMs speak for diverse people?Tuning LLMs via debate to generate controllable controversial statements.In Proceedings of the Association for Computational Linguistics:ACL 2024 (pp.16160-16176).Bangkok,Thailand:ACL.doi:10.18653/v1/2024.findings-acl.956.
- Matz,S.C.,Kosinski,M.,Nave,G.& Stillwell,D.J.(2017).Psychological targeting as an effective approach to digital mass persuasion.Proceedings of the National Academy of Sciences of the United States of America,114(48),12714-12719.doi:10.1073/pnas.1710966114.
- Matz,S.C.,Teeny,J.D.,Vaid,S.S.,Peters,H.,Harari,G.M.& Cerf,M.(2024).The potential of generative AI for personalized persuasion at scale.Scientific Reports,14(1),4692.doi:10.1038/S41598-024-53755-0.
- Mehrabi,N.,Morstatter,F.,Saxena,N.,Lerman,K.& Galstyan,A.(2022).A survey on bias and fairness in machine learning.ACM Computing Surveys (CSUR),54(6),115.doi:10.1145/3457607.
- Metaxa,D.,Park,J.S.,Robertson,R.E.,Karahalios,K.,Wilson,C.,Hancock,J.& Sandvig,C.(2021).Auditing algorithms:Understanding algorithmic systems from the outside in.Foundations and Trends?? in Human-Computer Interaction,14(4),272-344.doi:10.1561/1100000083.
- Micheli,M.,Lutz,C.& Büchi,M.(2018).Digital footprints:An emerging dimension of digital inequality.Journal of Information,Communication and Ethics in Society,16(3),242-251.doi:10.1108/jices-02-2018-0014.
- Noar,S.M.,Benac,C.N.& Harris,M.S.(2007).Does tailoring matter?Meta-analytic review of tailored print health behavior change interventions.Psychological Bulletin,133(4),673-693.doi:10.1037/0033-2909.133.4.673.
- Obermeyer,Z.,Powers,B.,Vogeli,C.& Mullainathan,S.(2019).Dissecting racial bias in an algorithm used to manage the health of populations.Science,366(6464),447-453.doi:10.1126/science.aax2342.
- Outrageous-Buy-9535.(2024).Fake sponsored ad in Advanced Voice Chat Convo [Reddit Post].r/ChatGPT.Retrieved from https://www.reddit.com/r/ChatGPT/commen ts/1fvq41q/fake_sponsored_ad_in_advanced_voice_chat_convo/
- Ouyang,L.,Wu,J.,Jiang,X.,Almeida,D.,Wainwright,C.L.,Mishkin,P.,Zhang,C.,Agarwal,S.,Slama,K.,Ray,A.,Schulman,J.,Hilton,J.,Kelton,F.,Miller,L.,Simens,M.,Askell,A.,Welinder,P.,Christiano,P.,Leike,J.& Lowe,R.(2022).Training language models to follow instructions with human feedback.In Proceedings of the 36th International Conference on Neural Information Processing Systems (pp.2011).New Orleans,LA,USA:Curran Associates Inc.
- Panickssery,A.,Bowman,S.R.& Feng,S.(2024).LLM evaluators recognize and favor their own generations.In Proceedings of the 38th International Conference on Neural Information Processing Systems (pp.2197).Vancouver,BC,Canada:Curran Associates Inc.
- Park,G.,Schwartz,H.A.,Eichstaedt,J.C.,Kern,M.L.,Kosinski,M.,Stillwell,D.J.,Ungar,L.H.& Seligman,M.E.P.(2015).Automatic personality assessment through social media language.Journal of Personality and Social Psychology,108(6),934-952.doi:10.1037/pspp0000020.
- Pasquale,F.(2016).The black box society:The secret algorithms that control money and information.Cambridge:Harvard University Press.
- Peters,H.& Matz,S.C.(2024).Large language models can infer psychological dispositions of social media users.PNAS Nexus,3(6),pgae231.doi:10.1093/pnasnexus/pgae231.
- Pit,P.,Ma,X.J.,Conway,M.,Chen,Q.Y.,Bailey,J.,Pit,H.,Keo,P.,Diep,W.& Jiang,Y.G.(2024).Whose side are you on?Investigating the political stance of large language models.arXiv preprint arXiv:2403.13840.doi:10.48550/arXiv.2403.13840.
- Sandvig,C.,Hamilton,K.,Karahalios,K.& Langbort,C.(2014).Auditing algorithms:Research methods for detecting discrimination on internet platforms.In Proceedings of the 64th Annual Meeting of the International Communication Association.Seattle,WA,USA.
- Santurkar,S.,Durmus,E.,Ladhak,F.,Lee,C.,Liang,P.& Hashimoto,T.(2023).Whose opinions do language models reflect?.In Proceedings of the 40th International Conference on Machine Learning (pp.1244).Honolulu,Hawaii,USA:JMLR.org.
- Schoenegger,P.,Salvi,F.,Liu,J.C.,Nan,X.L.,Debnath,R.,Fasolo,B.,Leivada,E.,Recchia,G.,Günther,F.,Zarifhonarvar,A.,Kwon,J.,Islam,Z.U.,Dehnert,M.,Lee,D.Y.H.,Reinecke,M.G.,Kamper,D.G.,Koba??,M.,Sandford,A.,Kgomo,J.,Hewitt,L.,Kapoor,S.,Oktar,K.,Kucuk,E.E.,Feng,B.,Jones,C.R.,Gainsburg,I.,Olschewski,S.,Heinzelmann,N.,Cruz,F.,Tappin,B.M.,Ma,T.,Park,P.S.,Onyonka,R.,Hjorth,A.,Slattery,P.,Zeng,Q.C.,Finke,L.,Grossmann,I.,Salatiello,A.& Karger,E.(2025).When large language models are more PersuasiveThan incentivized humans,and why.arXiv preprint arXiv:2505.09662.doi:10.48550/arXiv.2505.09662.
- Shi,W.& Li,J.H.(2024).New digital divide shaped by algorithm?Evidence from agent-based testing on Douyin's health-related video recommendation.Communication Research,51(7),867-890.doi:10.1177/00936502241262056.
- Stachl,C.,Au,Q.,Schoedel,R.,Gosling,S.D.,Harari,G.M.,Buschek,D.,V??lkel,S.T.,Schuwerk,T.,Oldemeier,M.,Ullmann,T.,Hussmann,H.,Bischl,B.& Bühner,M.(2020).Predicting personality from patterns of behavior collected with smartphones.Proceedings of the National Academy of Sciences of the United States of America,117(30),17680-17687.doi:10.1073/pnas.1920484117.
- Sundar,S.S.& Kim,J.(2019).Machine heuristic:When we trust computers more than humans with our personal information.In Proceedings of 2019 CHI Conference on Human Factors in Computing Systems (pp.538).Glasgow,Scotland,UK:ACM.doi:10.1145/3290605.3300768.
- Ta,N.,Zeng,J.& Li,Z.H.(2025).Governance of discriminatory content in conversational AIs:A cross-platform and cross-cultural analysis.Information,Communication & Society,1-18.doi:10.1080/1369118x.2025.2537803.
- Tappin,B.M.,Wittenberg,C.,Hewitt,L.B.,Berinsky,A.J.& Rand,D.G.(2023).Quantifying the potential persuasive returns to political microtargeting.Proceedings of the National Academy of Sciences of the United States of America,120(25),e2216261120.doi:10.1073/pnas.2216261120.
- Taubenfeld,A.,Dover,Y.,Reichart,R.& Goldstein,A.(2024).Systematic biases in LLM simulations of debates.In Proceedings of 2024 Conference on Empirical Methods in Natural Language Processing (pp.251-267).Miami,Florida,USA:ACL.doi:10.18653/v1/2024.emnlp-main.16.
- Teeny,J.D.& Matz,S.C.(2024).We need to understand “when” not “if” generative AI can enhance personalized persuasion.Proceedings of the National Academy of Sciences of the United States of America,121(43),e2418005121.doi:10.1073/pnas.2418005121.
- Weidinger,L.,Mellor,J.,Rauh,M.,Griffin,C.,Uesato,J.,Huang,P.S.,Cheng,M.,Glaese,M.,Balle,B.,Kasirzadeh,A.,Kenton,Z.,Brown,S.,Hawkins,W.,Stepleton,T.,Biles,C.,Birhane,A.,Haas,J.,Rimell,L.,Hendricks,L.A.,Isaac,W.,Legassick,S.,Irving,G.& Gabriel,I.(2021).Ethical and social risks of harm from Language Models.arXiv preprint arXiv:2112.04359.doi:10.48550/arXiv.2112.04359.
- Wu,Y.C.,Sun,E.,Zhu,K.J.,Lian,J.X.,Hernandez-Orallo,J.,Caliskan,A.& Wang,J.D.(2025).Personalized safety in LLMs:A benchmark and a planning-based agent approach.arXiv preprint arXiv:2505.18882.doi:10.48550/arXiv.2505.18882.
- Zhai,X.M.& Krajcik,J.(2022).Pseudo AI bias.arXiv preprint arXiv:2210.08141.doi:10.48550/arXiv.2210.08141.
- Zhao,X.Q.,Strasser,A.,Cappella,J.N.,Lerman,C.& Fishbein,M.(2011).A measure of perceived argument strength:Reliability and validity.Communication Methods and Measures,5(1),48-75.doi:10.1080/19312458.2010.547822.