Conversation-as-a-Database
a new framework where natural speech becomes structured input, enabling real-time, dynamic UI generation and intelligent task orchestration
The Interface is the Bottleneck
LLMs today can parse legal contracts, draft marketing campaigns, and write production-grade code. Yet, we still collect information and navigate tasks manually using static forms and dashboards. These rigid structures forces users to translate their rich, nuanced thoughts into a language machines understand.
We believe the next generation of AI tools will be able to natively understand our language. Beyond transcription, it would be able to structure it, plan around it, and build interactive experiences real time.
Today, we are coining a new term — Conversation-as-a-Database (CaaD). A new framework where natural speech becomes structured input, enabling real-time, dynamic UI generation and intelligent task orchestration.
What is Conversation-as-a-Database?
Conversation-as-a-Database (CaaD) is a new interface paradigm where your spoken words are treated as structured, queryable, and dynamic data. Instead of filling out forms or navigating dashboards, you simply talk.
As you speak, the system:
Captures your words and intent
Transforms them into structured data (like tables, tasks, dependencies)
Remembers context (short- and long-term)
Dynamically renders UI components in real time
It’s not voice dictation. It’s infrastructure that turns conversation into a living, editable, queryable dataset that drives intelligent action.
SQL was the lingo of data in the 1990s, and we believe conversation is the new schema of post-interface systems.
How It Works
Here’s an example of a flow, using gullie, a relocation voice agent that plans and executes your move:
Step 1: You speak naturally.
"I'm moving to Zurich in September with my partner and our dog. I need help figuring out visas, housing, and getting my pet there."
Step 2: The system parses this into structured data.
Entity: You
Destination: Zurich
Timeline: September
Stakeholders: Partner, Dog
Tasks: Visa prep, housing search, pet relocation
Step 3: A real-time personalized UI is generated.
Timeline widget shows countdown to September
Checklist of relocation tasks appears, pre-filled
Embedded cards suggest expat-friendly neighborhoods, pet travel rules, visa types
As the conversation continues, the system evolves:
"Actually, I think we’ll go in mid-October instead."
Timeline shifts. Deadlines and dependencies move. Related tasks are update.
This, is CaaD in action.
Why Now?
Three reasons make CaaD possible now:
LLM Maturity: Today’s models (like GPT-4o) can parse complex speech, handle context shifts, and structure information in near real-time.
Voice Infra Is Finally Ready: Tools like Whisper, ElevenLabs, and real-time streaming APIs make low-latency voice capture viable.
Dynamic UI Frameworks Are Commonplace: React, Vue, and others let developers build interfaces that update instantly based on data changes.
The final unlock is a shift in user behavior: people are ready to talk to software.
Designing with CaaD: New UX Principles
Designing for CaaD is fundamentally different:
From Fields to Flows: Inputs are created on demand, only when needed.
Time-aware and Adaptive: Interfaces evolve as the conversation does.
Context-Rich: Past mentions and historical conversations influence present suggestions.
Interruptible and Correctable: Users can change their mind mid-sentence and the system adapts.
Instead of designing UI first and building logic second, we reverse it: logic and memory first, then UI emerges dynamically.
Technical Stack Overview
To build CaaD systems, a few core layers are needed:
Voice Input Stack: High-fidelity, low-latency audio streaming (Whisper, Deepgram, ElevenLabs)
Semantic Parsing: LLMs to interpret, structure, and map intent
Context Memory: Short-term (per session) and long-term (across sessions)
UI Engine: Reactive frontend frameworks capable of real-time UI generation (React, Vue, Svelte)
Orchestration Layer: Logic engine to coordinate tasks, dependencies, and state updates
Challenges and Open Questions
CaaD is powerful, but early. Key questions include:
Data Ownership: How is conversational data stored, redacted, and governed?
Error Correction: How do users revise or undo misinterpreted inputs?
Trust & Transparency: How do we show users what’s been captured and why?
Latency: Can systems keep up with fast, casual human speech?
Design patterns are still forming. Best practices are emergent.
The Future is Voice-Native
We believe Conversation-as-a-Database will be the next foundational shift in how humans interact with machines. Interfaces will no longer be something we navigate. They will emerge as we speak.
In the near future:
You won’t fill out onboarding forms. You’ll describe your goals.
You won’t build dashboards. You’ll talk through scenarios.
You won’t Google things. You’ll ask, "can you just handle this for me?"
The tools that win will be those that listen better, remember more, and shape themselves to us.
Join Us
We’re pioneering this space and would love to collaborate with designers, engineers, and builders interested in redefining interaction itself.
If you’re building the next generation of voice-native systems, or want to co-develop parts of the CaaD stack, reach out. The interface is disappearing. What comes next is conversation.


