Published in HCII 2026 and CAAD Futures 2025 · Presented at BIG.AI @ MIT and McGill University AI Ethics Roundtable

/Beyond Renders

From Speculation to Specification - a compound AI system that grounds generative interior design images with real-world material data, transforming AI from visualization tool to decision-support partner.

Post-Generative GroundingCompound AISelf-Consistency SamplingMCP ServerResponsible AI
IIGenAI Demo - Workspace with Material Tiles
IIGenAI Demo - Workspace with Material Tiles
IIGenAI Demo - Material Detail Drawer Open
IIGenAI Demo - Mask Editing Mode Active
N=9
User Study
3
Conditions
< $0.03
Per Run
312
Materials

/The Expectation Gap

The Core Problem

Generative AI tools produce stunning interior renders in seconds. But a photorealistic ‘stone wall’ could be faux wallpaper, MDF panels, ceramic tiles, or actual stone masonry - each with different cost, carbon footprint, and structural requirements. The most creative phase of design is completely disconnected from the most impactful data.

“I conceptually want lower-carbon designs but when iterating with Midjourney I have zero material data.”
- Practicing architect, preliminary interviews
Three Paradigms
Pre-Generative

Prompt engineering, ControlNet, RAG - constrains exploration before the designer starts.

During-Generation

Fine-tuning, model modifications - requires 10K+ examples, weeks of GPU time, and produces static knowledge.

Post-Generative (IIGenAI)

Validate AFTER generation - preserves creative freedom, zero retraining, modular updates.

Visual Output + Domain Knowledge = Actionable Insights

/Three Computational Planes

Architecture

IIGenAI operates across three planes: the Interactive Plane captures user intent, the Inference Plane runs the AI pipeline, and the Grounding Plane validates against real-world databases.

Architecture Diagram - Three Computational Planes
Architecture Diagram - Three Computational Planes
Pipeline Flow - Five Step Process
Phase 1
Image Generation (gpt-image-1)+
Phase 2
Material Identification (5-Pass Self-Consistency)+
Phase 3
Chain-of-Thought Retry (o4-mini)+
Phase 4
CO₂e Grounding (ICE + Material2050)+
Phase 5
Confidence & Transparency UX+

/Research Impact

Behavioral Shift
5–10% → 100%

Sustainability consideration in design decisions. Baseline designers focused purely on aesthetics. With grounding, every participant explicitly addressed environmental impact using material-specific language like “rammed earth,” “mass timber,” and “low-carbon substitutes.”


Keyword shiftAesthetic → Material-specific
Creative Exploration
74 / 100

Creativity Support Index with grounding (vs 41.7 baseline). Counterintuitively, grounding INCREASED creative exploration. Designers used constraints deliberately and expressed intent more clearly. Satisfaction dropped to 50% - reflecting productive tension, not tool failure.


Prompt divergence0.92 → 0.83 cosine similarity
System Performance
< $0.03

Cost per full pipeline cycle. Image generation (~$0.015) + 5-pass VLM identification (~$0.008) + CoT retry (~$0.003) + grounding ($0). Compare: fine-tuning costs thousands and weeks of GPU time. The architecture is model-agnostic - swap any component independently.


ICE Database312 materials, 24 categories
“Satisfaction dropped because designers went from optimizing one thing - aesthetics - to balancing aesthetics against environmental impact. That tension is uncomfortable, and that's exactly what we want. The 50% tells us grounding is working.”

/Product Thinking

“IIGenAI is spell-check for buildability. It does not stop you from imagining - it tells you what is real.”

The Three Layer Product
Layer 1 - Generate (The Hook)

Designer types a prompt, gets a beautiful interior in seconds. This is table stakes - Midjourney and DALL-E already do this. The experience feels familiar and fast.

Layer 2 - Ground (The Value)

Every material in that image is automatically identified with confidence scores and CO₂e data from real engineering databases. This is what no other tool does. This is the moat.

Layer 3 - Iterate (The Retention Loop)

Designer sees concrete at high carbon, types “replace with rammed earth,” and the system edits the same image. The CO₂e drops. They learn, they explore, they come back.

Metrics as Product Validation
Activation100% engaged with sustainability data (vs 5–10% baseline)
Engagement DepthLexical diversity 0.38 → 0.42 - richer, more specific prompts
ExplorationPrompt divergence increased - more solution space explored
Creative SupportCSI rose from 41.7 → 74 - tool made designers MORE creative
Productive FrictionSatisfaction dropped to 50% - users hit real constraints for the first time
The Platform Play

The grounding layer is a platform, not a feature. Swap the database and you get a new product.

Architecture + ICE Database = Sustainable design tool
Museums + Provenance DB = Verified cultural curation
Product Design + Manufacturing DB = Feasibility checker
Film Production + Cost DB = Set design budgeting

The MCP server means any AI agent - including Claude - can plug into the grounding layer without rebuilding the pipeline.

/Vision & Next Steps

What I Learned

Post-generative grounding is a responsible AI strategy applicable far beyond architecture. The pattern - Visual Output + Domain Knowledge = Actionable Insights - generalizes to any creative domain where AI generates compelling outputs that lack domain-specific verification. The key insight is timing: validate after generation, not before. This preserves creative freedom while adding real-world accountability.

Built With

Claude Code (Sonnet 4.6) · Next.js 14 · FastAPI · OpenAI gpt-image-1 · GPT-4.1-mini · o4-mini · Anthropic MCP SDK · ICE Database V4.1 · Material2050 API

The Pattern Generalizes

Product design + manufacturing databases = tooling feasibility.
Marketing + brand guidelines = compliance scores.
Museums + provenance databases = verified cultural context.
E-commerce + sourcing APIs = supplier availability.

Future Vision
01
Canvas Workspace

Miro-style infinite canvas with SAM segmentation for automatic material region detection and spatial annotations.

02
Multi-Objective Optimization

Interactive sliders for CO₂e, cost, durability, acoustics - designers define their own utility function.

03
Product Recognition

Extend from materials to furniture and fixtures, linked to real supplier SKUs and product databases.

04
Multi-Modal Input

Sketch upload, voice commands, and 3D model integration for a synergistic design dialogue.

05
BIM Integration

Export grounded material data to Rhino/Revit, bridging generative AI and parametric design workflows.