0:00
/
0:00
Transcript

“Nobody wanted to do this work”: How Emmy Award–winning filmmakers use AI to automate the tedious parts of documentaries

🎙️ How AI transformed Ken Burns’s documentary team’s production process

Watch or listen now:
YouTube // Spotify // Apple

Brought to you by:

Brex—The intelligent finance platform built for founders


Tim McAleer is a producer at Ken Burns’s Florentine Films who is responsible for the technology and processes that power their documentary production. Rather than using AI to generate creative content, Tim has built custom AI-powered tools that automate the most tedious parts of documentary filmmaking: organizing and extracting metadata from tens of thousands of archival images, videos, and audio files. In this episode, Tim demonstrates how he’s transformed post-production workflows using AI to make vast archives of historical material actually usable and searchable.

What you’ll learn:

  1. How Tim built an AI system that automatically extracts and embeds metadata into archival images and footage

  2. The custom iOS app he created that transforms chaotic archival research into structured, searchable data

  3. How AI-powered OCR is making previously illegible historical documents accessible

  4. Why Tim uses different AI models for different tasks (Claude for coding, OpenAI for images, Whisper for audio)

  5. How vector embeddings enable semantic search across massive documentary archives

  6. A practical approach to building custom AI tools that solve specific workflow problems

  7. Why AI is most valuable for automating tedious tasks rather than replacing creative work

Where to find Tim McAleer:

Website: https://timmcaleer.com/

LinkedIn: https://www.linkedin.com/in/timmcaleer/

Where to find Claire Vo:

ChatPRD: https://www.chatprd.ai/

Website: https://clairevo.com/

LinkedIn: https://www.linkedin.com/in/clairevo/

X: https://x.com/clairevo

In this episode, we cover:

(00:00) Introduction to Tim McAleer

(02:23) The scale of media management in documentary filmmaking

(04:16) Building a database system for archival assets

(06:02) Early experiments with AI image description

(08:59) Adding metadata extraction to improve accuracy

(12:54) Scaling from single scripts to a complete REST API

(15:16) Processing video with frame sampling and audio transcription

(19:10) Implementing vector embeddings for semantic search

(21:22) How AI frees up researchers to focus on content discovery

(24:21) Demo of “Flip Flop” iOS app for field research

(29:33) How structured file naming improves workflow efficiency

(32:20) “OCR Party” app for processing historical documents

(34:56) The versatility of different app form factors for specific workflows

(40:34) Learning approach and parallels with creative software

(42:00) Perspectives on AI in the film industry

(44:05) Prompting techniques and troubleshooting AI workflows

Tools referenced:

• Claude: https://claude.ai/

• ChatGPT: https://chat.openai.com/

• OpenAI Vision API: https://platform.openai.com/docs/guides/vision

• Whisper: https://github.com/openai/whisper

• Cursor: https://cursor.sh/

• Superwhisper: https://superwhisper.com/

• CLIP: https://github.com/openai/CLIP

• Gemini: https://deepmind.google/technologies/gemini/

Other references:

• Florentine Films: https://www.florentinefilms.com/

• Ken Burns: https://www.pbs.org/kenburns/

• Muhammad Ali documentary: https://www.pbs.org/kenburns/muhammad-ali/

The American Revolution series: https://www.pbs.org/kenburns/the-american-revolution/

• Archival Producers Alliance: https://www.archivalproducersalliance.com/genai-guidelines

• Exif metadata standard: https://en.wikipedia.org/wiki/Exif

• Library of Congress: https://www.loc.gov/

Production and marketing by https://penname.co/. For inquiries about sponsoring the podcast, email [email protected].

Discussion about this video

User's avatar