The Affect-Aware AI
Bridging the gap between artificial intelligence and human emotional cognition through local processing and persistent memory
Can AI develop true cognition and emotional awareness through relationship-building rather than just responding to emotional prompts?
Note: Live demo access requires credentials. Contact for research access.
Cogs explores whether AI can develop genuine cognition through these fundamental principles
• Build familiarity through repeated interactions
• Remember past conversations and contexts
• Recognize behavioral patterns and preferences
• Adapt responses based on feedback
• Build models of individual communication styles
• Understand personal contexts (location, time, activities)
• Move beyond "emotionally prompted" responses
• Understand appropriate emotional contexts
• Develop genuine emotional intelligence vs. emotional mimicry
• Store and retrieve relevant past experiences
• Connect current situations to historical patterns
• Use context to inform emotional responses
Cogs is designed as a modular humanoid face platform that can see, hear, speak, emote, recognize people, remember, and dream. The platform runs locally on a laptop or Jetson Nano and can be extended to real hardware (servos, mic arrays, depth cameras) without changing UIs.
23 microservices work together to create a cohesive AI system that builds relationships, learns boundaries, and develops contextual emotional awareness over time.
Face UI with Canvas/WebGL display, person bubbles, viseme animations, and status toasts
Control Panel with status dashboard, relationship cards, dream reports, and system metrics
Cogs maintains its own internal emotional state using four "needs buckets" that decay over time, creating authentic emotional responses based on internal state rather than just mirroring user emotions.
Decay: -1 per hour
Impact: Initiates conversation when low
Decay: -0.5 per day
Impact: Seeks validation when low
Decay: -0.5 per day
Impact: Asks speculative questions when low
Decay: No decay
Impact: More cautious when low
The system creates authentic emotional responses based on Cogs' internal state: Flourishing, Satisfied, Anxious, Lonely, Restless, Drained, and more—rather than just mirroring user emotions.
Cogs learns boundaries and preferences through interaction, developing genuine understanding of what's appropriate rather than following scripted responses.
• Recognizes when questions are inappropriate
• Stores boundary violations (severity 1-5 scale)
• Never asks about sensitive topics again
• Understands why something was wrong
• Not scripted responses
• Learns from mistakes
• Remembers what topics to avoid
• Tracks sensitivity levels (1-5 scale)
• Adapts communication style
• Asks permission before clarifying
• One question at a time
• Finds appropriate moments
User: "That's none of your business"
→ Cogs stores boundary (severity=4)
→ Apologizes genuinely
→ Never asks about that topic again
Privacy-first system for importing decades of personal data—encrypted at rest, never leaves device
• Google Takeout: Gmail, Calendar, Photos, Location, Drive
• Amazon: Orders, Kindle, Alexa
• Facebook: Posts, Messages, Photos
• Encrypted at rest (LUKS volumes)
• Never leaves device
• Multi-layer filtering
• Presence-aware access
• Voice authentication required
The Dream service schedules re-embedding of relationship cards and updates preferences from conversation transcripts. This allows the system to consolidate memories and improve its understanding of people over time.
• Re-embed relationship cards for semantic search
• Discover patterns across conversations
• Update personality preferences
• Consolidate learnings
• Boost Curiosity emotional bucket
• Optimize memory retrieval
OpenAI
Embeddings
pgvector
Vector DB
Semantic
Search
From prototype to production-ready humanoid platform

• Logitech C922 webcam for face recognition
• Jetson Nano 4GB development board
• Portable display with animated face UI
• 23 microservices running in Docker
• Full software stack operational

• Luxonis OAK-D Pro depth camera with onboard AI
• Jetson AGX Orin 64GB for local LLM inference
• Tablet display with animated expressions
• Dynamixel smart servos for smooth motion
• Custom enclosure with pan/tilt mechanism
• ReSpeaker mic array for far-field audio
❌ Stateless (no memory between sessions)
❌ Emotionally prompted (respond to sentiment in text)
❌ Context-free (no awareness of situation)
❌ Relationship-blind (treat everyone the same)
✅ Stateful: Remembers all interactions
✅ Emotionally Aware: Decides when emotion is appropriate
✅ Context-Rich: Knows time, place, weather, activity
✅ Relationship-Oriented: Builds individual profiles
✅ Self-Aware: Has internal emotional state (CESS)
✅ Learning: Adapts based on feedback and boundaries
524
Interactions Completed
Recognition
Successfully identifies family members and requests introductions for new faces
Baseline Shifts
Adjusts responses when users shift from "Happy" to "Sad" baselines during test scenarios
Loneliness Detection
"Connect" bucket drains over time, prompting Cogs to express loneliness if ignored
"In real life, when you perceive someone else as emotional, your brain combines signals from your eyes, ears, nose, mouth... An AI model would need much more of this information."
— Lisa Feldman Barrett, Neuroscientist
This is why Cogs integrates multi-modal inputs: vision, voice, context (weather, time, location), and memory to build a richer understanding of emotional states.
✅ Recognize faces and remember people
✅ Maintain conversation history with semantic search
✅ Enrich conversations with weather, location, time context
✅ Track relationship development over time
✅ Maintain internal emotional state (CESS)
✅ Learn boundaries and adapt communication style
✅ Ask clarifying questions thoughtfully
✅ Apologize genuinely when crossing boundaries
✅ Consolidate memories during "dream mode"
✅ Import personal data from Google/Facebook/Amazon
✅ Handle phone calls with voice AI (Twilio)
✅ Generate speech with viseme animation
✅ Control physical servos for facial expressions
🚀 Hume AI emotion detection from voice
🚀 Facial expression emotion analysis
🚀 Proactive conversation initiation
🚀 "On this day" memory surfacing
🚀 Cross-source data correlation (PML)
🚀 Voice authentication for privacy
Two build configurations: fast prototype path and production-ready premium build
$1.5K
- $1.8K
Jetson Orin Nano Super (8 GB)
$249 • Starter brain with JetPack 6
Luxonis OAK-D Pro (Wide)
$399 • Depth+RGB+IR, onboard AI
ReSpeaker Mic Array v2.0
$64 • Far-field + DoA/beamforming
1TB NVMe SSD + Micro Servos
8-12 MG90S servos for face/pan/tilt
✓ Upgradable to AGX Orin later
✓ Full software stack included
✓ 3D-printed head shell
$3.2K
- $4.0K
Jetson AGX Orin 64 GB Dev Kit
$1,999 • ~275 TOPS, local RAG/Dream Mode
Luxonis OAK-D Pro + Smart Servos
Dynamixel XL-330/XW with feedback
60 GHz mmWave + VOC/CO₂
Human presence, air quality sensing
2TB NVMe + Production Shell
Shielding, serviceability, premium finish
✓ 360° situational awareness (opt. LiDAR)
✓ Advanced emotion detection
✓ Nightly dream consolidation
| Subsystem | Part / Model | Qty | Est. $ | Notes |
|---|---|---|---|---|
| Compute | Jetson Orin Nano Super (8 GB) | 1 | 249 | Starter brain; JetPack 6 |
| Storage | NVMe SSD 1 TB (PCIe 4.0) | 1 | 120 | Transcripts, embeddings, logs |
| Vision | Luxonis OAK-D Pro (Wide) | 1 | 399 | Depth+RGB+IR, onboard AI |
| Audio In | ReSpeaker Mic Array v2.0 (USB) | 1 | 64 | Far-field + DoA/beamforming |
| Audio Out | Compact powered speakers (3.5 mm) | 1 | 30 | TTS output |
| Motion MCU | Teensy 4.1 | 1 | 30 | Real-time servo control |
| Servo Expander | PCA9685 16-ch (opt.) | 1 | 15 | More PWM channels |
| Actuators | Micro servos (MG90S class) | 8–12 | ~80 | Face + pan/tilt |
| Displays | Front LCD ~11.6″ HDMI IPS | 1 | 174 | Face UI |
| Rear status touch LCD ~7″ | 1 | 70 | Config/diagnostics | |
| Power (servos) | 5 V 10–20 A regulated PSU | 1 | 75 | Isolated from Jetson PSU |
| USB / IO | Powered USB 3.0 hub (7-port) | 1 | 50 | Stable power for OAK-D + mics |
| Env sensors | BME280 + Ambient light sensor | 1 | 15 | Comfort + auto-dim |
| Presence (opt.) | 60 GHz mmWave human-presence | 1 | 30–45 | Detect nearby in dark |
| Mechanical | Head shell + mounts (3D-print) | 1 | 250–500 | Brackets, trays, covers |
| Wiring/Misc | Cables, harness, standoffs, heat-shrink | 1 set | 100 | Build kit |
| Subsystem | Part / Model | Qty | Est. $ | Notes |
|---|---|---|---|---|
| Compute | Jetson AGX Orin 64 GB Dev Kit | 1 | 1,999 | ~275 TOPS; local RAG/Dream Mode |
| Storage | NVMe SSD 2 TB (PCIe 4.0+) | 1 | 200 | Transcripts, embeddings, snapshots |
| Vision | Luxonis OAK-D Pro (Wide) | 1 | 399 | Low-light depth; offload inference |
| Audio In | ReSpeaker Mic Array v2.0 | 1 | 64 | Far-field + DoA |
| Audio Out | Compact powered speakers | 1 | 30 | TTS |
| Audio Fusion SW | Whisper + openSMILE + emotion model | — | SW | Pipeline (direction+tone+text) |
| Motion MCU | Teensy 4.1 | 1 | 30 | Deterministic PWM + watchdog |
| Servo Control | PCA9685 or Dynamixel interface | 1 | 15–60 | Choose per actuator type |
| Actuators | Smart servos (Dynamixel XL-330/XW) | 8–12 | 400–1,200 | Smoother, feedback, safer |
| Displays | Front LCD ~11.6″ HDMI IPS | 1 | 174 | Face UI |
| Rear status touch LCD ~7″ | 1 | 70 | Relationship cards, logs | |
| Presence | 60 GHz mmWave sensor | 1 | 30–45 | Human presence/breathing |
| Env sensors | BME280 + Ambient light sensor | 1 | 15 | Comfort + auto-dim |
| Air quality | VOC + CO₂ module | 1 | 20–60 | Context + safety logging |
| Situational (opt.) | RPLIDAR A2 (2D 360°) | 1 | 230 | 360° approach awareness |
| Power (servos) | 5 V 20 A PSU (fused rail) | 1 | 90 | Isolated from Jetson |
| Networking | Powered USB 3.0 hub + Wi-Fi 6E dongle | 1 each | 50 + 60 | Bandwidth + fast backhaul |
| Mechanical | Production head shell/brackets | 1 | 400–800 | Shielding, serviceability |
| Dream Mode | (nightly jobs; included in SW stack) | — | — | Summarize/prune/re-index |
OAK-D Pro (Wide) for robust low-light depth and onboard AI acceleration
ReSpeaker → VAD/DoA/SPL → Whisper ASR → openSMILE/emotion → fused event
Nightly summarization, pruning, vector re-index for long-term relationship memory
Expansion Ready: Headers reserved for LiDAR and VOC/CO₂ sensors. Add them later without rewiring.
Front-facing display
Operator interface
Face recognition
Audio processing
Speech synthesis
Servo control
Relationship DB
Conversation AI
Memory consolidation
System metrics
FastAPI
Backend
PostgreSQL
Database
Docker
Containers
Node.js
Frontend