Submit questions and messages on Speakp">

Latent Space: The AI Engineer Podcast

Agents @ Work: Dust.tt

Overview

  • Stanislas Polu's journey spans from traditional tech roles at Oracle and Stripe to pioneering mathematical reasoning research at OpenAI, where he worked closely with Ilya Sutskever before departing in 2022 to found Dust, sensing a pivotal moment before potential AGI emergence.
  • Dust evolved from a developer-focused open-source tool to a platform enabling non-technical users to create personalized AI assistants that automate workflows through "programming with English" rather than complex API integration, achieving impressive 88% daily active user rates in some enterprise deployments.
  • The platform addresses critical infrastructure challenges rather than model limitations, focusing on creating "pipes" for agents to access data and take actions across systems like Notion, Slack, and GitHub, similar to Stripe's infrastructure approach.
  • Polu envisions AI fundamentally transforming traditional SaaS by creating more direct interaction interfaces, potentially enabling smaller teams to achieve significant scale and shifting human roles toward strategic decision-making and curation rather than execution.

Content: Stanislas Polu's AI Journey and Dust Platform

Background and Early Career

  • Podcast features Stanislas Polu, co-founder and CEO of Dust, discussing his AI journey (part of a two-part series on AI agents and productivity)
  • Educational background: École Polytechnique and Stanford
  • Professional experience: Oracle, Stripe, and OpenAI
  • Connected to early OpenAI team members like Greg Brockman and Daniela Amodei
  • First AI exposure was at Stanford in an early machine learning class taught by Andrew Ng
  • Fascinated by computer science since age 16

Path to OpenAI

  • After Stripe, explored potential AI impact areas:
- Self-driving cars (decided against due to safety concerns) - Cybersecurity (attempted transformer-based fuzzing, which didn't succeed) - Mathematical applications of AI
  • Connected with OpenAI through Greg and then Ilya Sutskever
  • Initially faced challenges joining as a remote employee from Paris, ultimately started as a contractor
  • Not a formally trained researcher (no PhD), but joined through connections and demonstrated engineering skills

Research at OpenAI

  • Focused on mathematical reasoning capabilities of AI:
- Explored combination of transformer creativity and formal mathematical system verification - Worked on mathematical problem-solving, particularly low-end math benchmarks like AMC - Treated proofs as programs with verification capabilities
  • Worked closely with Ilya Sutskever, who served as an advisor and coach
  • Most compute at OpenAI was reserved for training GPTs, with teams exploring specific problems
  • Compute allocation was a key management tool to guide research priorities
  • Fine-tuned over 10,000 models using 10 million A100 hours
  • Research evolved from GPT-2 towards more advanced models

OpenAI Culture and Leadership

  • Ilya Sutskever provided the "North Star" vision for the organization
- Strongly believed in scaling compute and the "compression thesis" - Focused on communicating vision and getting teams to work together towards AGI
  • Impressed by Sam Altman's ability to:
- Quickly understand complex technical details despite having no ML background - Switch between high-level CEO perspective and deep technical understanding - Stay informed about ground-level work while maintaining strategic overview

OpenAI Split and Departure

  • Present during the OpenAI split but wasn't deeply involved in the details
  • Key uncertainties about the split included:
- Potential disagreements over technology commercialization - Possible differences around API development - Potential safety concerns
  • OpenAI recovered quickly due to clear organizational mission, computational resources, and talent
  • Transitioned from OpenAI in September 2022
  • Saw founding a company as potentially the "last train leaving the station" before potential AGI emergence
  • Motivated to create something more product-focused after experiencing research work
  • Viewed GPT-4's potential as a clear signal to start a company

Founding Dust

  • Initially developed Dust as an open-source tool for developers working with LLMs
  • Early product development focused on:
- Creating tools for developers, not enterprises - Building workflows requiring multiple examples to avoid overfitting - Developing a UI to help developers introspect and observe LLM interactions
  • Open source strategy driven by:
- Transparency with security teams - Ability to show code and development progress - Engaging with technical users - Facilitating bug hunting

Dust's Evolution and Mission

  • Evolved from a technical platform alternative to Langchain
  • Now an open-source platform for building personalized AI assistants
  • Core mission:
- Build infrastructure for companies to deploy AI agents within teams - Enable non-developers (curious "tinkerers") to create operational agents - Focus on infrastructure and product development
  • Infrastructure approach:
- Create "pipes" for agents to access data and take actions - Maintain connections to platforms like Notion, Slack, GitHub - Comparable to Stripe's infrastructure value

Agent Capabilities and Design Philosophy

  • Current agent capabilities focus on simple workflows involving multiple tools
  • "Programming with English" - describing tasks without complex API compatibility
  • Tools can include semantic search, database queries, web searches, external actions
  • Function calling works best with precise, scripted instructions
  • Challenges emerge when providing high-level, broad instructions with many tools
  • Potential in creating simple agents that can then be used as actions for more complex "meta agents"

Integration Strategy

  • Primarily uses APIs for system integrations, avoiding browser automation when possible
  • For Salesforce integration, exploring a model where users provide specific queries to inject data
  • Most SaaS products used by target companies (500-5,000 employee tech companies) now have APIs
  • Current web integration methods are "broken", primarily using search and headless browsing
  • Excited about emerging technologies that can render web pages in model-compatible formats and expose web actions to AI models

Real-World Applications

  • Creating AI assistants to automate parts of weekly meetings:
- Generating tables from Slack data - Creating financial graphs and reports
  • Long-term goal is a comprehensive "weekly meeting assistant" that can run automatically
  • Currently focused on simple, individual agents rather than complex "meta agents"
  • Deliberately targeting non-technical users with accessible language and design

Company Philosophy and Approach

  • Product-focused company, not a research company
  • Uses existing AI models rather than training their own
  • Aims to push product boundaries with current model capabilities
  • Challenges in evaluating agent performance and interaction quality
  • Proposed solutions include using models to evaluate conversations and building product features for user feedback

Model Selection and Performance

  • Companies may need to own their models through internal post-training realignment loops
  • Model selection becoming more abstracted for non-technical users
  • Function calling considered critically important for AI interactions
  • GPT-4 Turbo currently perceived as potentially better at function calling than GPT-4.0

Enterprise Adoption and Results

  • Achieved high user penetration in enterprise settings:
- 88% daily active users in some companies - 60-70% weekly active users on average
  • Focus on creating significant product improvements rather than marginal model performance gains

Infrastructure Challenges

  • Current AI limitations are more about infrastructure than model quality
  • Integrating data from different platforms is complex and requires nuanced handling
  • Different data types (tabular, text) require distinct processing approaches
  • Vertical infrastructure and orchestration are critical
  • Uses Temporal as an orchestration tool for managing complex workflows
  • Handles asynchronous tasks like scheduling and conditional execution

Future of SaaS and AI

  • High-growth companies traditionally prefer "buying" solutions over building
  • Post-high-growth companies are now reconsidering SaaS investments
  • AI may reduce need for traditional SaaS product layers by creating more direct interaction interfaces
  • Software might evolve into a "labor interface" where AI/agents perform tasks
  • Potential for smaller teams (around 20 people) to achieve significant scale with AI assistance

Content Creation and AI

  • Discussed potential billionaire content creators (Mr. Beast, Joe Rogan)
  • Human creators serve as editors, making key decisions
  • Example: Daily AI newsletter where human selects and curates content
  • CEO/creator's primary function becomes making strategic "yes/no" choices

Product Development Strategy

  • Vertical vs. horizontal AI approach:
- Horizontal creates broad, company-wide tools (significant value but harder go-to-market) - Vertical offers easier market entry (solving specific industry problems)
  • Recommendation: Build specific product in one niche
- Use product success to drive potential future platform development - Allows for brand establishment and potential lateral expansion

More from Latent Space: The AI Engineer Podcast

Explore all episode briefs from this podcast

View All Episodes →

Listen smarter with PodBrief

Get AI-powered briefs for all your favorite podcasts, plus a daily feed that keeps you informed.

Download on the App Store