Instabase founder and CEO Anant Bhardwaj joins a16z Infra partner Guido Appenzeller to discuss the revolutionary impact of LLMs on analyzing uns">

AI + a16z

Giving New Life to Unstructured Data with LLMs and Agents

Key Takeaways

Deep Dive

The Unstructured Data Challenge and Early Research

The conversation begins with defining the core problem: unstructured data - information that cannot be easily placed into database tables or processed with SQL queries, including PDFs, images, and various document formats. Traditional processing methods were severely limited and brittle, relying on:

Anant Bardwaj's journey started at MIT in 2015, where he began researching unstructured data management through a project called Data Hub, exploring ways to ask questions about heterogeneous, unstructured data. He later moved to Silicon Valley to explore commercial applications, recognizing that many enterprises struggle with processing unstructured data for critical decision-making processes like immigration visa applications and loan approvals.

The technological landscape at the time was characterized by rudimentary techniques, with Robotic Process Automation (RPA) showing significant limitations when dealing with unstructured data. The vision was that AI and large language models would drive substantial automation improvements through more decentralized, federated execution of data processing.

Technical Innovation and Breakthrough Development

The research and development journey initially focused on program synthesis approaches to solve unstructured data problems, experimenting with regular expressions and program generation. Early experiments with BERT and transformer models yielded poor results, leading to a novel innovation.

The breakthrough came with InstaLM, a revolutionary model that encoded tokens with both position in sentence and X-Y coordinates. This BERT-like model could understand document layout, achieving significant improvements in document understanding and helping triple company revenue from 2021 to 2022.

When OpenAI's ChatGPT launched in November 2022, it initially seemed potentially disruptive. However, the team realized that LLMs alone are not a complete solution, recognizing the need for compound AI systems with multiple components. This led to understanding that "size matters" in AI model performance, and LLMs could enable advanced document processing, sorting, and applications in banking and insurance for document analysis.

Complex Document Processing and System Design

The discussion moves to real-world challenges, particularly home loan applications that are often 100+ pages with complex unstructured documents. Traditional document processing suffers from reliability and completeness issues, while LLMs can make surprising errors, especially with complex documents due to context window limitations and potential missed details. Simply using RAG (Retrieval-Augmented Generation) proves insufficient for critical enterprise applications.

The proposed solution approach involves developing comprehensive workflows for document processing using:

Key enterprise use cases include reducing lending processing time from weeks to seconds, systematic document review for intelligence analysis and threat detection, and comprehensive document parsing with guaranteed completeness. The fundamental insight emerges that AI system design should focus on building explainable, auditable workflows that recognize AI systems are not 100% perfect, implementing human escalation paths and designing systems with reasonable error rates and verification mechanisms.

Enterprise Reliability and Error Management

A significant shift in enterprise thinking becomes apparent: companies are moving from seeking 100% accuracy to prioritizing predictability. Acceptable error rates are emerging, with focus on detecting when errors occur, understanding error nature, and creating systems to minimize error impact. Unpredictable errors prove more problematic than occasional errors, leading enterprises to want systems that can identify which portions need human review and provide transparency about potential inaccuracies.

The future of information processing involves AI transforming document and information handling by generating summaries, pre-parsing complex documents, highlighting key points of interest, and reducing information to essential elements. An innovative example emerges: a bank in India offering lending services entirely through WhatsApp, demonstrating conversational AI enabling new customer interaction models and radically different user experiences, particularly transformative in developing markets.

AI Transformation and Enterprise Adoption Challenges

AI's potential to fundamentally change user experiences extends across various business processes, with improvements in call centers, account opening, lending, document processing, and immigration applications. However, enterprise adoption faces two primary barriers: data safety and security concerns, and auditability and predictability of AI decisions.

Key enterprise requirements include transparent AI decision-making processes, ability to explain and trace AI steps similar to human workflows, consistent runtime behavior, and compliance with internal regulations. Emerging trends point toward AI agents as a new user interface paradigm, shifting from step-by-step transactions to high-level autonomous instructions, with potential for multi-agent systems making collaborative decisions.

Current challenges with AI agents include lack of deterministic path selection and inconsistent runtime behavior when given the same goals and tools, while enterprises prefer predictable, consistent processes.

Practical AI Implementation and Future Vision

The conversation reveals a crucial insight: AI agents are most valuable during build/compile time, not runtime. AI can produce first drafts of code, workflows, and control paths that humans can review and refine, while runtime processes should remain deterministic, auditable, and debuggable. Fully autonomous AI systems are not yet practical, suggesting a more pragmatic approach of "freezing" AI-generated workflows once they work, with enterprise AI being controlled similar to how companies manage employee decision-making.

The future vision for AI execution considers two potential models:

Development focuses on a "federated AI execution" framework where thousands of agents can dynamically discover and communicate with each other, collaborate on complex tasks without central control, and share capabilities while determining optimal execution paths.

Technical Implementation and Industry Evolution

The discussion addresses evolution from RPA to AI-driven approaches, with focus on solving unstructured data problems for automation. AI is positioned to potentially replace RPA by enabling more dynamic system interactions. Key technical developments include introducing the Model Context Protocol (MCP) for dynamically discovering and calling system capabilities, exploring end-to-end workflows using AI agents during compile time, and proposing "identity pass-through" for managing user permissions and system access.

Current limitations encompass authentication, system compatibility, and handling potential failures, emphasizing the importance of setting constraints on AI agent capabilities during initial setup and maintaining human control by defining runtime behavior during compile time. The philosophical approach treats AI agents like "interns" with controlled capabilities, separating compile-time configuration from runtime execution to maintain human oversight.

Strategic Business Imperative

The conversation concludes with a critical business perspective: technological shifts require early adoption by enterprises, despite potential complications. The risks of not adapting include potential business obsolescence, using Barnes & Noble as a cautionary example.

Enterprises can benefit from new technologies through more efficient workflows, improved customer and partner experiences, significant cost savings, accelerated operational processes, and transformative customer experiences. The critical question shifts from whether to adopt new technologies to how to successfully implement them, with confidence in the potential and necessity of technological adaptation for business survival and growth.

More from AI + a16z

Explore all episode briefs from this podcast

View All Episodes →

Listen smarter with PodBrief

Get AI-powered briefs for all your favorite podcasts, plus a daily feed that keeps you informed.

Download on the App Store