人机协作决策的效能异质性:聚焦主体性的四阶段过程影响模型*

耿晓伟, 李鑫琦, 徐志平, 谢天

心理科学 ›› 2025, Vol. 48 ›› Issue (4) : 933-947.

PDF(683 KB)
中文  |  English
PDF(683 KB)
心理科学 ›› 2025, Vol. 48 ›› Issue (4) : 933-947. DOI: 10.16719/j.cnki.1671-6981.20250414
计算建模与人工智能

人机协作决策的效能异质性:聚焦主体性的四阶段过程影响模型*

  • 耿晓伟1,2, 李鑫琦1,2, 徐志平1,2, 谢天**3,4
作者信息 +

The Heterogeneity of the Effectiveness of Human-AI Collaborative Decision-Making: A Four-Stage Process Influence Model Focusing on Agency

  • Geng Xiaowei1,2, Li Xinqi1,2, Xu Zhiping1,2, Xie Tian3,4
Author information +
文章历史 +

摘要

人机协作决策在各领域广泛应用,但其效能表现出显著异质性——既可能提升也可能降低决策质量与效率。为解释效能异质性的来源,本研究基于人机共生理论视角,聚焦人类主体性,构建了一个人机协作决策的四阶段过程影响模型。该模型揭示了人机协作效能形成于四个相互关联的动态环节:(1)人机优劣势与偏见的识别;(2)基于任务特性与情境的动态任务分配;(3)对AI建议质量的辨别与信任校准;(4)合理人机依赖模式的形成与调整。在深入探讨了人机协作面临的核心挑战,并重点剖析了人类主体性与决策能力弱化的风险后,本研究展望了未来构建人机共享认知以实现深度协同的前景,从而为理解和优化人机协作决策提供了新的理论框架与思考维度。

Abstract

The integration of artificial intelligence (AI) into the decision-making has revolutionized fields such as healthcare, finance, and criminal justice by offering the potential for enhanced efficiency and accuracy through human-AI collaboration decision-making (HAIC-DM). However, empirical outcomes remain inconsistent. While AI augments human capabilities in complex tasks (e.g., AI-assisted medical diagnostics matching expert performance), it can also degrade performance in simpler tasks due to cognitive redundancy or overreliance (e.g., automation bias in image recognition). This heterogeneity stems from unresolved tensions between technological potential and human factors. First, misaligned task allocation often undermines complementary strengths. While AI excels at structured, data-driven tasks (e.g., credit scoring), it has limitations in contextual reasoning and ethical judgment necessitate human oversight—a balance frequently disrupted in practice. Second, asymmetric trust dynamics skew collaboration: opaque AI systems (e.g., "black-box" algorithms) foster overreliance or distrust, as seen in radiologists uncritically accepting erroneous high-confidence AI diagnoses. Third, bias amplification—where algorithmic biases (e.g., racial disparities in recidivism prediction tools) intersect with human cognitive heuristics (e.g., anchoring effects), creating self-reinforcing error cycles that exacerbate inequities in judicial and hiring decisions. The urgency to reconcile AI’s computational power with human agency, particularly in ethically sensitive contexts, underscores the need for systematic exploration of collaborative mechanisms and risks.
This study synthesizes 54 empirical studies from computer science, psychology, and organizational research (2018-2024). These studies were retrieved from the ACM Digital Library, the Web of Science, and the AIS eLibrary, using keywords such as "human-AI collaboration" and "decision-making." The inclusion criteria prioritized quantitative assessments of HAIC-DM performance (human-only, AI-only, and collaborative outcomes). A thematic analysis was conducted to identify recurring patterns in task characteristics (e.g., structured vs. unstructured goals), interaction designs (e.g., explanation formats), and moderators (e.g., user expertise). A four-stage Process Impact Model was developed, integrating principles from symbiosis theory and distributed cognition. Case studies (e.g., healthcare diagnostics, autonomous driving) were analyzed to validate stage-specific mechanisms, and experimental findings (e.g., trust calibration experiments) informed theoretical refinements.
The proposed model identifies four interdependent stages that govern the efficacy of HAIC-DM:
(1)Strengths/Biases Recognition: AI excels at structured tasks (e.g., fraud detection), while humans dominate ethical judgments. Biases, such as algorithmic (e.g., historical data biases) and cognitive biases (e.g., anchoring effects), distort collaboration in judicial decisions where humans and AI redundantly overemphasize prior convictions.
(2)Context-Driven Task Allocation: Optimal allocation improves accuracy (e.g., AI pre-screening cancer images + human validation boosts diagnostic accuracy by 15%), whereas misallocation (e.g., AI-led creative writing) yields superficial outputs.
(3)Trust Calibration: Example-based explanations improve the discernment of advice (+22% accuracy in income prediction), yet opaque AI systems induce overreliance, as radiologists often accept erroneous high-confidence diagnoses.
(4)Adaptive Dependency: Balanced reliance maximizes efficacy (e.g., AI risk alerts in autonomous driving + human ethical oversight), but over-dependence triggers cognitive offloading and eroding skills (e.g., lawyers who rely excessively on AI for contract analysis).
This study advances HAIC-DM research by framing collaboration as a co-evolutionary process. It emphasizes bidirectional adaptation between humans (critical thinking, ethical oversight) and AI (transparency, contextual learning). The Process Impact Model clarifies how dynamic interactions, from bias recognition to dependency calibration, determine efficacy. It offers actionable insights for optimizing task allocation and trust mechanisms. Future work must prioritize shared mental models to align AI’s computational logic with human intuition, particularly in high-stakes domains like healthcare and criminal justice. Institutional reforms, including ethical governance frameworks and mandatory human oversight protocols, are critical to mitigate risks like accountability erosion. Fostering synergistic interdependence, where AI augments human cognition without supplanting agency, is key to realizing the vision of "humans as ethical navigators, AI as precision enablers." This alignment ensures that collaborative intelligence enhances, rather than undermines, societal decision-making in an AI-augmented future.

关键词

人机协作决策 / 决策效能 / 效能异质性 / 人机共生 / 主体性 / 过程影响模型

Key words

human-AI collaboration / human-AI symbiosis / decision efficacy / agency / process model / shared cognition

引用本文

导出引用
耿晓伟, 李鑫琦, 徐志平, 谢天. 人机协作决策的效能异质性:聚焦主体性的四阶段过程影响模型*[J]. 心理科学. 2025, 48(4): 933-947 https://doi.org/10.16719/j.cnki.1671-6981.20250414
Geng Xiaowei, Li Xinqi, Xu Zhiping, Xie Tian. The Heterogeneity of the Effectiveness of Human-AI Collaborative Decision-Making: A Four-Stage Process Influence Model Focusing on Agency[J]. Journal of Psychological Science. 2025, 48(4): 933-947 https://doi.org/10.16719/j.cnki.1671-6981.20250414

参考文献

[1] 吴刚, 赵军, 苏静逸, 谢鑫. (2022). “工作-学习”理论的创新与发展——第四代“文化—历史”活动理论及应用价值. 远程教育杂志, 40(2), 86-95.
[2] 笛卡尔. (1986). 第一哲学沉思集 (庞景仁译). 商务印书馆.
[3] 韩敏, 赵海明. (2020). 智能时代身体主体性的颠覆与重构——兼论人类与人工智能的主体间性. 西南民族大学学报(人文社科版), 41(5), 56-63.
[4] 康德. (2004). 纯粹理性批判 (邓晓芒译). 人民出版社..
[5] 康德. (2010). 实践理性批判 (邓晓芒译). 人民出版社..
[6] 涂良川, 乔良. (2021). 人工智能“高阶自动化”的主体可能性——兼论人工智能奇点论的存在论追问. 现代哲学, 6, 32-40.
[7] 殷杰. (2024). 生成式人工智能的主体性问题. 中国社会科学, 8, 124-145+207.
[8] 于雪, 翟文静, 侯茂鑫. (2022). 人工智能时代人机共生的模式及其演化特征探究. 科学与社会, 12(4), 106-119.
[9] 朱奥, 刘华硕, 袁佩君, 张丹. (2025). 迈向人机共生时代: 人机协同决策的心理学视角. 科技导报, 43(3), 37-46.
[10] Baier P., DeLallo D., & Sviokla J. J. (2024). Your organization isn' t designed to work with GenAI. Harvard Business Review.
[11] Bansal G., Nushi B., Kamar E., Weld D. S., Lasecki W. S., & Horvitz, E. (2019). Updates in human-AI teams: Understanding and addressing the performance/compatibility tradeoff. Proceedings of the 33rd AAAI Conference on artificial intelligence and 31st innovative applications of artificial intelligence conference and 9th aaai symposium on educational advances in artificial intelligence, AAAI Press.
[12] Bansal G., Wu T., Zhou J., Fok R., Nushi B., Kamar E., Ribeiro M. T., & Weld D. (2021). Does the whole exceed its parts? The effect of AI explanations on complementary team performance. Proceedings of the 2021 CHI conference on human factors in computing systems.
[13] Buçinca Z., Lin P., Gajos K. Z., & Glassman E. L. (2020). Proxy tasks and subjective measures can be misleading in evaluating explainable AI systems. Proceedings of the 25th international conference on intelligent user interfaces.
[14] Buçinca Z., Malaya M. B., & Gajos K. Z. (2021). To trust or to think: Cognitive forcing functions can reduce overreliance on AI in AI-assisted decision-making. Proceedings of the ACM on Human-Computer Interaction, 5, 188.
[15] Chen V., Liao Q. V., Wortman Vaughan J., & Bansal G. (2023). Understanding the role of human intuition on reliance in human-AI decision-making with explanations. Proceedings of the ACM on Human-Computer Interaction, 7, 370.
[16] Christman, J. (2020). Autonomy in moral and political philosophy. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab, Stanford University.
[17] Cox, A. (2024). Algorithmic literacy, ai literacy and responsible generative AI literacy. Journal of Web Librarianship, 18(3), 93-110.
[18] Eshleman, A. (2024). Moral responsibility. In E. N. Zalta & U. Nodelman (Eds.), The Stanford encyclopedia of philosophy. Metaphysics Research Lab, Stanford University.
[19] Green, B., & Chen, Y. (2019). Disparate interactions: An algorithm-in-the-loop analysis of fairness in risk assessments. In Proceedings of the conference on fairness, accountability, and transparency.
[20] Guo Z., Wu Y., Hartline J. D., & Hullman J. (2024). A decision theoretic framework for measuring AI reliance. Proceedings of the 2024 ACM conference on fairness, accountability, and transparency.
[21] He Z., Song Y., Zhou S., & Cai Z. (2023). Interaction of thoughts: Towards mediating task assignment in human-AI cooperation with a capability-aware shared mental model. Proceedings of the 2023 CHI conference on human factors in computing systems.
[22] Hu T., Kyrychenko Y., Rathje S., Collier N., van der Linden S., & Roozenbeek J. (2025). Generative language models exhibit social identity biases. Nature Computational Science, 5, 65-75.
[23] Hutchins, E. (1995). Cognition in the wild. The MIT Press..
[24] Jacobs M., Pradier M. F. Mccoy. T. H., Perlis R. H., Doshi-Velez F., Gajos. K. Z. (2021). How machine-learning recommendations influence clinician treatment selections: The example of antidepressant selection. Translational Psychiatry, 11, 108.
[25] Jo E., Epstein D. A., Jung H., & Kim Y.H. (2023). Understanding the benefits and challenges of deploying conversational AI leveraging large language models for public health intervention. Proceedings of the 2023 CHI conference on human factors in computing systems.
[26] Jussupow E., Spohrer K., Heinzl A., & Gawlitza J. (2021). Augmenting medical diagnosis decisions? An investigation into physicians' decision-making process with artificial intelligence. Information Systems Research, 32(3), 713-735.
[27] Krügel, S., & Uhl, M. (2024). The risk ethics of autonomous vehicles: An empirical approach. Scientific Reports, 14, 960.
[28] Kuutti, K. (1995). Activity theory as a potential framework for human-computer interaction research. In B. Nardi (Ed.), Context and consciousness: Activity theory and human-computer interaction (pp. 17-44). MIT Press.
[29] Lai V., Liu H., & Tan C. (2020). Why is "Chicago" deceptive? Towards building model-driven tutorials for humans. Proceedings of the 2020 CHI conference on human factors in computing system.
[30] Lai, V., & Tan, C. (2019). On human predictions with explanations and predictions of machine learning models: A case study on deception detection. Proceedings of the conference on fairness, accountability, and transparency.
[31] Liang G., Sloane J. F., Donkin C., & Newell B. R. (2022). Adapting to the algorithm: How accuracy comparisons promote the use of a decision aid. Cognitive Research: Principles and Implications, 7(1), 14.
[32] Licklider, J. C. R. (1960). Man-computer symbiosis. IRE Transactions on Human Factors in Electronics, 1(1), 4-11.
[33] Liu H., Lai V., & Tan C. (2021). Understanding the effect of out-of-distribution examples and interactive explanations on human-AI decision making. Proceedings of the ACM on Human-Computer Interaction, 5, 408.
[34] Liu, X. B., (2023). Experiencing visual saptions: Augmented communication with real-time visuals using large language models. Adjunct Proceedings of the 36th annual ACM symposium on user interface software and technology.
[35] Maadi M., Akbarzadeh Khorshidi H., & Aickelin U. (2021). A review on human-AI interaction in machine learning and insights for medical applications. International Journal of Environmental Research and Public Health, 18(4), 2121.
[36] Mirowski P., Mathewson K. W., Pittman J., & Evans R. (2023). Co-writing screenplays and theatre scripts with language models: Evaluation by industry professionals. Proceedings of the 2023 CHI conference on human factors in computing systems.
[37] Nicolescu, L., & Tudorache, M. T. (2022). Human-computer interaction in customer service: The experience with AI chatbots—A systematic literature review. Electronics, 11, 1579.
[38] Noti, G., & Chen, Y. (2022). Learning when to advise human decision makers. Proceedings of the 33rd international joint conference on artificial intelligence.
[39] Noy, S., & Zhang, W. (2023). Experimental evidence on the productivity effects of generative artificial intelligence. Science, 381, 187-192.
[40] Papenmeier A., Kern D., Englebienne G., & Seifert C. (2022). It’s complicated: The relationship between user trust, model accuracy and explanations in AI. ACM Transactions on Computer-Human Interaction, 29(4), Article 34.
[41] Parasuraman, R., & Riley, V. (1997). Humans and automation: Use, misuse, disuse, abuse. Human Factors, 39(2), 230-253.
[42] Petridis S., Diakopoulos N., Crowston. K., Hansen M., Henderson K., Jastrzebski S., Nickerson J., & Chilton. L. B. (2023). AngleKindling: Supporting journalistic angle ideation with large language models. Proceedings of the 2023 CHI conference on human factors in computing systems .
[43] Rastogi C., Zhang Y., Wei D., Varshney K. R., Dhurandhar A., & Tomsett R. (2022). Deciding fast and slow: The role of cognitive biases in AI-assisted decision-making. Proceedings of the ACM on Human-Computer Interaction.
[44] Rastogi C., Leqi L., Holstein K., & Heidari H. (2023). A taxonomy of human and ML strengths in decision-making to investigate human-ML complementarity. Proceedings of the AAAI conference on human computation and crowdsourcing, 11, 127-139.
[45] Reverberi C., Rigon T., Solari. A., Hassan. C., Cherubini P., & Cherubini A. (2022). Experimental evidence of effective human-AI collaboration in medical decision-making. Scientific Reports, 12, 14952.
[46] Schemmer M., Hemmer P., Nitsche M., Kühl N., & Vössing M. (2022). A meta-analysis of the utility of explainable artificial intelligence in human-AI decision-making. Proceedings of the 2022 AAAI/ACM Conference on AI, ethics, and society.
[47] Shin W., Han J., & Rhee W. (2021). AI-assistance for predictive maintenance of renewable energy systems. Energy, 221, 119775.
[48] Shneiderman, B. (2020). Human-centered artificial intelligence: Reliable, safe & trustworthy. International Journal of Human-Computer Interaction, 36(6), 495-504.
[49] Tschandl P., Rinner. C., Apalla Z., Argenziano. G., Codella N., Halpern A., & Kitter H. (2020). Human-computer collaboration for skin cancer recognition. Nature Medicine, 26, 1229-1234.
[50] Vaccaro M., Almaatouq A., & Malone T. (2024). When combinations of humans and AI are useful: A systematic review and meta-analysis. Nature Human Behaviour, 8(12), 2293-2303.
[51] Vaccaro, M., & Waldo, J. (2019). The effects of mixing machine learning and human judgment. Communications of the ACM, 62(11), 104-110.
[52] Vasconcelos, H., et al. (2023). Explanations can reduce overreliance on AI systems during decision-making. Proceedings of the ACM on Human-Computer Interaction, 7, 129.
[53] Vimpari V., Kultima A., Hämäläinen P., & Guckelsberger C. (2023). 'An adapt-or-die type of situation': Perception, adoption, and use of text-to-image-generation AI by game industry professionals. Proceedings of the ACM on Human-Computer Interaction, 7, 131-164.
[54] Wilcox L., Brewer R., & Diaz F. (2023). AI consent futures: A case study on voice data collection with clinicians. Proceedings of the ACM on Human-Computer Interaction, 7, 316.
[55] Wilder B., Horvitz E., & Kamar E. (2020). Learning to complement humans. Proceedings of the 29th International Joint Conference on Artificial Intelligence.
[56] Wilson, H. J., & Daugherty, P. R. (2018). Collaborative intelligence: Humans and AI are joining forces. Harvard Business Review, 96, 114-123.
[57] Wilson H. J.,& Daugherty, P. R. (2018). Human + machine: Reimagining work in the age of AI. Harvard Business Review Press.
[58] Zhang Y., Liao Q. V., & Bellamy, R. K. E. (2020). Effect of confidence and explanation on accuracy and trust calibration in AI-assisted decision making. Proceedings of the 2020 conference on fairness, accountability, and transparency.

基金

*本研究得到浙江省高校重大人文社科攻关计划项目(2024GH005)和教育部人文社会科学重点研究基地重大项目(22JJD860013)的资助

PDF(683 KB)

评审附件

Accesses

Citation

Detail

段落导航
相关文章

/