In recent years, the integration of artificial intelligence into educational settings has sparked intense debate concerning its efficacy and the impact on student engagement and learning. A pioneering study led by Dr. Joshua Lambert, an associate professor and biostatistician at the University of Cincinnati College of Nursing, delved into the role of AI-powered chatbots in higher education, specifically within nursing education. His investigation revolves around understanding whether these AI tools can augment learning experiences by providing reliable, interactive, and non-judgmental academic support to students pursuing advanced degrees.
Dr. Lambert’s research entailed a pilot study wherein Doctor of Nursing Practice (DNP) students interacted with a custom-designed AI chatbot, evaluating its performance against responses from human educators—a professor and a graduate assistant. The design of the study was meticulous, employing a blinded, randomized, within-subjects comparison approach to minimize bias. This meant that the participating students were unaware of the origin of each response they evaluated, facilitating an objective assessment based solely on content quality, helpfulness, and overall satisfaction.
The study’s methodology was grounded in survey-based data collection, with seven doctoral students submitting statistical questions relevant to their capstone work. Each participant received three separate responses: one from a professor, one from a graduate assistant, and one generated by the AI chatbot. The students then rated each answer using a five-point Likert scale across dimensions of helpfulness, satisfaction, and hypothetical future utility. Moreover, they were tasked with guessing which response originated from the chatbot, providing insightful data on user perceptions and biases towards AI.
Intriguingly, the chatbot’s responses achieved the highest ratings in terms of both satisfaction and helpfulness. This outcome challenges prevailing assumptions about the inferiority of AI-generated academic assistance compared to human input. However, the findings also uncovered a paradox; when students attempted to identify which responses were from the chatbot, they frequently misattributed the lowest-rated responses as AI-generated. This suggests a cognitive bias wherein skepticism or distrust towards AI platforms colors students’ judgment, despite acknowledging their efficacy.
Dr. Lambert interprets this skepticism as a reflection of the broader issue of trust in AI adoption across academia. The dissonance between preferring AI responses yet mistrusting their source reveals an underlying psychological barrier to seamless AI integration. This duality emphasizes the necessity for educational strategies that not only optimize AI tools’ functionality but also foster trust and acceptance among both learners and educators.
The study’s nuanced insights align with contemporary research indicating that user trust is a pivotal determinant in AI’s widespread acceptance and efficacy. Particularly in academic environments, where the stakes of knowledge accuracy and intellectual development are high, trust in AI must be cultivated deliberately. Transparency in AI operations, consistent accuracy, and clear communication about the chatbot’s limitations and strengths are critical components in this trust-building process.
Collaborators in this investigation included distinguished faculty members Robyn Stamm, Shannon White, and Melanie Kroger-Jarvis from the University of Cincinnati College of Nursing, alongside Dr. Bailey Martin from the University of Colorado Anschutz Medical Campus. Together, they underscore the interdisciplinary and multi-institutional commitment to advancing knowledge on AI’s pedagogical applications. Their shared conclusion advocates for larger-scale, multisite studies incorporating qualitative and quantitative analyses to fully elucidate AI’s role in education.
While this pilot study serves as an essential proof-of-concept within nursing education, its implications extend far beyond. The potential for AI chatbots to lower social and psychological barriers for students—especially when posing questions they might hesitate to ask human instructors due to fear of judgment or appearing uninformed—is transformative. AI offers a non-judgmental, accessible, and immediate resource empowering students to engage more deeply and confidently with challenging material.
Moreover, Dr. Lambert highlights that these AI tools could mitigate the intimidation students sometimes feel in traditional academic settings. The chatbot acts as a knowledge consultant devoid of bias or impatience, fostering an inclusive environment where intellectual curiosity is unhindered by social anxieties. This attribute is particularly valuable in rigorous disciplines such as nursing, where mastering complex statistical and clinical concepts is essential.
However, the researchers caution that despite promising preliminary results, the small sample size of seven students restricts the generalizability of conclusions. It is vital to avoid overextending interpretations beyond initial findings. Future studies with expanded cohorts and diverse demographics will be crucial to validate the efficacy and relevance of AI chatbots as sustainable educational tools.
Funding support for this exploratory research came from the University of Cincinnati College of Nursing, alleviating costs related to conference attendance, participant incentives, and software licensing. The research team reports no conflicts of interest, ensuring impartiality in the study’s design and analysis. Their transparent acknowledgment of study limitations and ethical rigor bolsters the credibility of this emerging scholarly dialogue around AI in education.
Ultimately, this study illuminates a complex, evolving landscape where chatbot technology intersects with human cognition, educational psychology, and institutional practices. It urges educators, technologists, and policymakers to collaboratively navigate the challenges of AI integration, emphasizing both performance metrics and the intangible factors such as trust and acceptance that will dictate the future trajectory of learning innovation. The University of Cincinnati’s work thereby marks a significant step toward harnessing AI to enhance student success, engagement, and autonomy in higher education contexts.
Subject of Research: People
Article Title: Blinded But Biased: Students Prefer Chatbot Until They Know It Is One
News Publication Date: 1-Apr-2026
References:
Lambert, J., Stamm, R., White, S., Kroger-Jarvis, M., & Martin, B. (2026). Blinded But Biased: Students Prefer Chatbot Until They Know It Is One. Journal of Nursing Education. https://journals.healio.com/doi/10.3928/01484834-20260216-01
Image Credits: Photo provided by the University of Cincinnati
Keywords: AI chatbot, higher education, nursing education, student satisfaction, artificial intelligence, trust in AI, educational technology, biostatistics, Doctor of Nursing Practice, user bias

