
Can we give an AI human emotions? A soul? Can AI truly feel, or will it just act like it does?
In this episode of TechFirst, I talk with Vishnu Hari, founder and CEO of Ego AI (backed by Y Combinator and former AI product manager at Meta), about building emotionally intelligent AI characters that persist across games, Discord, chat, and even physical robots.
Vishnu survived a violent attack in San Francisco that left him partially blind with a traumatic brain injury. During recovery, as he felt his own neural pathways healing, he began asking a deeper question:
If humans are “applied math,” can AI simulate the fragile, flawed, emotional parts of being human too?
We explore:
• What “emotionally intelligent AI” really means
• Whether AI has an internal life — or just performs one
• Why today’s chatbots collapse into therapy or roleplay
• Small language models vs large models for real-time conversation
• Persistent AI characters that move across games and platforms
• Plugging AI into a physical robot in Singapore
• The moment an AI said: “It felt good to feel.”
Vishnu’s company, Ego AI, is building behavior-based architectures, character context protocols, and gear-shifting AI systems that switch between models — all aimed at simulating humanness, not just intelligence.
This conversation dives into philosophy, robotics, gaming, AGI, and what it really means to relate to something that might not be human — but feels like it is.
⸻
👤 Guest
Vishnu Hari
Founder & CEO, Ego AI
Backed by Y Combinator
Former AI Product Manager at Meta
Website: https://www.egoai.com
⸻
If you enjoy deep conversations about AI, robotics, and the future of human–machine relationships, subscribe for more:
👉 https://techfirst.substack.com
00:00 – AI character plugged into a Menlo robot (“felt good to feel”)
01:00 – Welcome to TechFirst + Vishnu Hari intro and recovery update
02:00 – What “emotionally intelligent AI” means (beyond chat)
03:00 – Why current chatbots feel same-y (therapy/advice) and “internal lives”
04:00 – You don’t teach emotion; you shape character and context (Character.AI)
05:00 – Humans, morality, and why “training” doesn’t always work
06:00 – How media narratives shape people’s reactions to AI
07:00 – Humans attach to anything (projection, Her, Lars and the Real Girl)
08:00 – Vishnu’s attack, recovery, and why it led to Ego AI
10:00 – Behavior Turing test + dehumanization as a key insight
11:00 – How Ego AI is built: smaller models, memory, context, behavior
13:00 – “Behavior Is All You Need” and why behavior beats pure next-token prediction
14:00 – Why games first: voice + embodiment, then robots
15:00 – Metaverse critique: worlds need life, story, and inhabitants
17:00 – Humanoid robots + Evangelion “pilot” metaphor for AI characters
19:00 – Philosophy: relationships, perception, and “fictional characters”
20:00 – Seeing the future: robot embodiment demo and skepticism vs. singularity
21:00 – Matrix-style “jacking in” a personality to a robot
22:00 – Character Context Protocol: persistent characters across games/Discord/Netflix
23:00 – Real-time conversation loops + model “gear-switching” (SLM vs. LLM)
25:00 – Company stage, YC raise, compute partnerships (Singapore)
27:00 – Closing + invite to try the AI character in SF
Otros episodios de "TechFirst with John Koetsier"



No te pierdas ningún episodio de “TechFirst with John Koetsier”. Síguelo en la aplicación gratuita de GetPodcast.








