Podcli ships as a Model Context Protocol server with 19 tools that cover transcription, AI clip scoring, face-tracked 9:16 cropping, burned-in captions, and export. Plug it into Claude Code, Claude Desktop, or Codex. Say “clip this episode” and the agent runs the whole pipeline.
MCP is the spec that lets an AI agent discover and call tools on your machine. Podcli runs as one of those servers.
Every step of the podcast-to-shorts pipeline is exposed as a tool. The agent picks the right ones automatically.
The MCP config is one block. Same shape for Claude Code, Claude Desktop, and Codex.
// claude_desktop_config.json (same shape for Codex) { "mcpServers": { "podcli": { "command": "node", "args": ["dist/index.js"], "cwd": "/path/to/podcli" } } }
Natural language. The agent reads the tool list and figures out the call sequence.
Same prompt, two tool surfaces. The agent does less work when the tools speak podcast, not pixels.
The questions people search before they wire Podcli into their MCP client.
A Model Context Protocol server that exposes 19 tools covering the podcast-to-clip pipeline: transcription, AI clip scoring against a knowledge base, face-tracked 9:16 cropping, burned-in captions via Remotion, hardware-accelerated export.
Install Podcli (git clone, ./setup.sh), then point your MCP client config at the built server. The same JSON works for Claude Desktop and Codex.
No. It runs locally as a process the agent talks to. Video, transcript, and renders all stay on your machine. The only optional network calls are to the Claude or Codex API for AI clip scoring.
Yes. The CLI is a fixed command. The MCP server exposes discrete tools the agent composes on demand. Both ship in the same install and you can use either.
Any MCP-compatible client. Tested with Claude Code, Claude Desktop, and Codex. Any agent that speaks MCP can read the tool list and call it.
Clone the repo, run the setup script, then point your MCP client at the built server. Takes about as long as making coffee.