Grounds Kaltura Avatars messaging in science: media richness, CASA paradigm, self-disclosure, anthropomorphism, uncanny valley, and pedagogical agent effects. Use these citations when prospects push back on "why avatars" or when exec audiences want proof, not hype.
- Why this matters
- The anchor: Alan Dennis (Kelley School of Business, Indiana University)
- Computers Are Social Actors (Nass and Reeves)
- Self-disclosure: people tell AI things they won't tell humans
- Anthropomorphism: how humanlike is enough
- Uncanny valley: the real constraint
- Pedagogical agents and learning
- Comparison: video agents vs text chatbots
- Healthcare and customer support
- How to use this in sales conversations
- Open gaps worth flagging internally
- Source library (for citations)
- Positioning one-liners grounded in the research
Why this matters
Prospects want proof before they believe the story. "Avatars sound cool" is not a purchase reason. This note collects the actual academic work that supports the three claims we keep making in pitches:
- Rich, face-to-face-like media beats text for complex interactions.
- People trust and disclose to AI agents differently than to humans, sometimes more.
- Humanlike cues (voice, face, expressions) change behavior even when users know the agent is a machine.
Use the references below as grounding when you write RFP responses, business cases, or executive emails. When a buyer asks "where is the evidence?", you now have 40 years of IS, HCI, and marketing research to point at.
The anchor: Alan Dennis (Kelley School of Business, Indiana University)
Dennis is the most useful academic to name when you want to ground the Kaltura Avatars story. He has 150+ papers and has moved from the foundational communication theories of the 1980s-90s directly into digital humans today. Three building blocks.
Media Richness Theory (Daft and Lengel, 1986)
Not Dennis's own theory, but the starting point. Daft and Lengel proposed that media vary in their ability to carry rich cues: face-to-face is richest (voice, facial expression, body language, instant feedback), written text is leanest. For equivocal or ambiguous tasks (negotiation, persuasion, emotional content, complex decisions), richer media produce better outcomes. For routine information transfer, leaner media are fine.
Implication for Kaltura Avatars: text chatbots are a lean-media tool solving a rich-media problem. Any task that involves trust, nuance, or decision-making belongs in a richer medium. Avatars give enterprises that richness at scale.
Media Synchronicity Theory (Dennis and Valacich, 1999)
Dennis's own extension, published at HICSS 1999 and later refined in MIS Quarterly. It replaced "richness" with five concrete media capabilities: synchronicity, parallelism, reprocessability, rehearsability, and symbol sets. Communication is split into two processes: conveyance (transferring information) and convergence (reaching shared understanding). Lean media work for conveyance, synchronous rich media work for convergence.
Implication: conversational avatars are optimized for convergence tasks - the kind that drive deals, adoption, onboarding, and training. Chat widgets are optimized for conveyance. Different tools, different jobs.
Less Artificial, More Intelligent: Understanding Affinity, Trustworthiness, and Preference for Digital Humans
Published in Information Systems Research, 2025 (vol. 36, issue 2, pp. 1096-1128). This is the Dennis paper to cite when somebody asks "is there actual research on digital humans?"
Key finding: digital human agents controlled by advanced AI produce higher user affinity, higher perceived trustworthiness, higher preference, and higher purchase likelihood than both text chatbots and traditional point-and-click interfaces. The jump is not marginal.
AI Recommendations Amplify Social Influence
Published in Journal of Management Information Systems, 2025. Demonstrates that when AI is labeled as the source of a recommendation, the recommendation itself carries more persuasive weight, not less. Counterintuitive and useful: saying "our AI avatar recommends..." can outperform "our team recommends...".
NVIDIA GTC 2025 session
Dennis's GTC talk (S72988) frames the telco customer experience thesis: digital humans replace chatbots as the primary conversational interface, unifying voice, face, and intent handling in one agent. Worth watching end-to-end and quoting from when pitching telco, BFSI, or any service-heavy vertical.
Computers Are Social Actors (Nass and Reeves)
Foundational paradigm from Clifford Nass and Byron Reeves at Stanford, built out across 30+ experiments in the 1990s and consolidated in The Media Equation (1996). Core finding: people apply social scripts to computers automatically, even when they deny doing so. Politeness, gender stereotypes, flattery effects, reciprocity, personality attribution, they all transfer.
For Kaltura Avatars the takeaway is precise: adding voice and a face does not just "feel nicer", it triggers measurably different behavior. Users are more polite, more patient, more forgiving after errors, and more likely to attribute expertise. The MASA extension (Media Are Social Actors, 2020+) updates the framework for today's avatars and agents: primary social cues (voice, face, name, conversational turn-taking) are enough to activate social treatment.
Use this when a CIO asks "why not just a better chatbot". Answer: because decades of evidence show that the presence of a face and voice is not a skin on top of a chatbot, it changes the interaction itself.
Self-disclosure: people tell AI things they won't tell humans
The most counterintuitive and most useful body of research for our pitch.
Lucas and Gratch (USC Institute for Creative Technologies, 2014)
Paper: "It's only a computer: Virtual humans increase willingness to disclose", Computers in Human Behavior, 2014. Platform: SimSensei, virtual human named Ellie.
239 participants. Same Ellie interviewer in all conditions. Half were told she was autonomous AI ("computer frame"), half were told she was teleoperated by a clinician ("human frame"). Participants in the computer frame disclosed more personal and embarrassing information, showed more intense sadness when discussing difficult topics, and were rated by observers as more open. Perceived absence of human judgment removed the impression-management barrier.
Why this matters for Kaltura: prospects worry avatars will "feel cold". The research says the opposite for the behaviors we care about. Users share more with a non-judgmental AI agent than with a human, particularly on sensitive topics. HR onboarding, compliance training, healthcare intake, employee listening, customer feedback - all of these benefit.
Broader Gratch and ICT work
Gratch, Lucas, Marsella, and colleagues at USC ICT have 15+ years of work on empathetic virtual humans, including EMA and GRETA platforms for modeling emotion and nonverbal behavior. The through-line: virtual humans that display contingent nonverbal feedback (head nods, gaze, expressions) produce disclosure and rapport at or above human-interviewer levels.
Anthropomorphism: how humanlike is enough
Blut et al. (2021) meta-analysis
Published in Journal of the Academy of Marketing Science, 2021. Synthesized 108 independent samples and 11,053 participants across physical robots, chatbots, and AI agents. This is the go-to number when a prospect wants "the big meta-analysis".
Headline: anthropomorphism positively predicts intention to use AI, mediated by perceived intelligence and usefulness. Effects are moderated by robot gender, service type, and customer traits. Anthropomorphism is not a single lever, it is a design space - cues must match task.
Practical synthesis from HCI work (2023-2025)
High anthropomorphism in chatbot avatars raises perceived empathy (β = 0.32) and trust (β = 0.27), which in turn drive user experience (β = 0.48, p < 0.01). Direct effects are small; the action is in the mediation. Translation: the face does not "make people buy", it makes them feel heard, which makes them buy.
Humanlike AI agents also change interaction style: users greet them, say please and thank you, accept errors more gracefully, and treat them as collaborators rather than tools. Workplace studies (Cambridge Judge Business School, 2025) show this shifts the tone from transactional to relational.
Uncanny valley: the real constraint
Mori (1970) first proposed the curve. MacDorman and colleagues (2009) empirically validated the drop in comfort and trust when realism is high but imperfect, especially in motion.
Consequences for avatar design:
- Low-realism, stylized avatars are safe. Comfort scales with likeability, not realism.
- High-realism avatars work only if motion, gaze, lip-sync, and micro-expressions are coherent. Otherwise users feel eerie and trust drops hard.
- The middle of the curve is dangerous. "Almost human" is worse than "clearly stylized".
Useful when a customer asks "why do your avatars look the way they look?". Answer: intentional design choice, validated by 50 years of research. We optimize for coherence across all cues, not for pure visual realism.
Pedagogical agents and learning
When the pitch lands in L and D, corporate training, or higher education, this is the relevant evidence base.
Schroeder et al. (2013) and Castro-Alonso et al. (2021)
Meta-analyses show small-to-moderate positive effects of pedagogical agents on learning performance versus no-agent conditions. 2D animated agents often beat 3D agents. Animated beats static. Effects are stronger on retention, motivation, and social presence than on raw knowledge tests.
Recent work (2024, Frontiers in Education)
Humanoid AI agents as teachers, with learner avatars, positively predicted performance, satisfaction, attention, and cognitive presence. Extends the meta-analytic finding into AI-driven agents.
Bottom line for training use cases: avatars do not magically raise test scores, but they reliably raise attention, completion, and perceived social presence, which are the levers that matter in corporate L and D where disengagement is the real enemy.
Comparison: video agents vs text chatbots
Synthesized from the sources above, useful as a one-slide summary.
| Dimension | Text chatbot | Video agent / digital human |
|---|---|---|
| Media richness | Lean | Rich (voice, face, gaze, timing) |
| Best use | Conveyance, routine tasks | Convergence, trust, emotion, decision |
| Social treatment | Minimal | Full social script activation (CASA) |
| Disclosure | Moderate | Higher when framed as non-judgmental AI |
| Purchase intent | Baseline | Higher affinity, trust, preference (Dennis 2025) |
| Learning use | Low engagement | Measurable lift in attention and retention |
| Risk | Dull, abandoned | Uncanny valley if realism is mismatched |
Healthcare and customer support
Studies on conversational agents in healthcare consistently show higher patient satisfaction with video than with voice-only (86% vs 77%), higher disclosure on sensitive topics (per Lucas and Gratch), and better comprehension when an agent pauses, rephrases, and responds to confusion cues. Applies directly to our pitches in insurance, pharma, and any post-sale support scenario.
How to use this in sales conversations
Short list of moves, mapped to buyer type.
- Skeptical CIO: cite Dennis 2025 (Information Systems Research) and Blut 2021 meta-analysis. "This is not a demo, this is 40 years of IS research."
- Head of CX: cite Nass and Reeves + MASA. "Users apply social scripts to faces and voices automatically. Text UIs leave that value on the table."
- HR, L and D, or compliance lead: cite Lucas and Gratch 2014. "People disclose more to non-judgmental AI interviewers. Better signal for you."
- Marketing leader: cite Dennis 2025 on AI-labeled recommendations amplifying persuasion. "An AI avatar saying 'I recommend X' carries more weight than a team page saying it."
- Design-minded buyer worried about uncanny valley: cite Mori 1970 and MacDorman 2009. "We design for cue coherence, not pure realism. That is why our avatars look the way they do."
Open gaps worth flagging internally
- No strong peer-reviewed B2B conversion rate study yet for AI avatars. Vendor claims exist (UBS 5x video volume, BHuman reply-rate lifts) but lack control groups. Opportunity for Kaltura to run and publish a rigorous study.
- Effect-size data in Blut 2021 requires full paper access. Get it for the library.
- Dennis's full post-2023 bibliography is wider than what the public abstracts show. Worth a direct outreach if we want a co-authored case study.
Source library (for citations)
- Daft, R.L., Lengel, R.H. (1986). Organizational Information Requirements, Media Richness and Structural Design. Management Science.
- Dennis, A.R., Valacich, J.S. (1999). Rethinking Media Richness: Towards a Theory of Media Synchronicity. HICSS-32. Later refined in MIS Quarterly (2008).
- Dennis, A.R. et al. (2025). Less Artificial, More Intelligent: Understanding Affinity, Trustworthiness, and Preference for Digital Humans. Information Systems Research, 36(2), 1096-1128. https://ideas.repec.org/a/inm/orisre/v36y2025i2p1096-1128.html
- Dennis, A.R. et al. (2025). Artificial Intelligence Recommendations Amplify the... Journal of Management Information Systems.
- Reeves, B., Nass, C. (1996). The Media Equation. Cambridge University Press / CSLI.
- Nass, C., Moon, Y. (2000). Machines and Mindlessness: Social Responses to Computers. Journal of Social Issues.
- Lucas, G.M., Gratch, J., King, A., Morency, L.P. (2014). It's only a computer: Virtual humans increase willingness to disclose. Computers in Human Behavior.
- Blut, M., Wang, C., Wünderlich, N.V., Brock, C. (2021). Understanding Anthropomorphism in Service Provision: A Meta-analysis. Journal of the Academy of Marketing Science, 49, 632-658.
- Mori, M. (1970, translated 2012). The Uncanny Valley. IEEE Robotics and Automation.
- MacDorman, K.F., Green, R.D., Ho, C.C., Koch, C.T. (2009). Too real for comfort? Uncanny responses to computer generated faces. Computers in Human Behavior.
- Schroeder, N.L., Adesope, O.O., Gilbert, R.B. (2013). How Effective are Pedagogical Agents for Learning? Journal of Educational Computing Research.
- Castro-Alonso, J.C. et al. (2021). Meta-analysis of pedagogical agents.
- Frontiers in Education (2024). Humanoid AI agents as teachers with learner avatars.
- NVIDIA GTC 2025 S72988. AI Agents and Digital Humans Shaping the Future of Interaction in Telecoms. Alan Dennis. https://www.nvidia.com/en-us/on-demand/session/gtc25-s72988/
Positioning one-liners grounded in the research
Copy-paste ready for decks, emails, RFPs.
- "Forty years of information systems research say that face, voice, and turn-taking change how people make decisions. Text chatbots were built for the opposite problem."
- "Users trust an AI agent differently than a human. In healthcare studies they disclosed more to a virtual interviewer when they believed it was fully automated. That is a feature, not a bug, for onboarding, training, and feedback."
- "The latest peer-reviewed research (Dennis, 2025, Information Systems Research) finds digital human agents produce higher affinity, higher trustworthiness, and higher purchase intent than chatbots or point-and-click UIs. We did not invent that curve, we built a product for it."
- "People apply social rules to computers with a face and a voice, automatically. Our avatars turn that reflex into engagement, at enterprise scale."