claude-video-vision
MCP server that gives Claude Code the ability to watch and understand videos — extracts frames via ffmpeg and processes audio via multiple backends
AI & MLnpx -y claude-video-vision{
"mcpServers": {
"claude-video-vision": {
"command": "npx",
"args": [
"-y",
"claude-video-vision"
]
}
}
}claude-video-vision is a community MCP server that connects AI assistants like Claude to mcp server that gives claude code the ability to watch and understand videos — extracts frames via ffmpeg and processes audio via multiple backends. It runs locally on your machine, keeping your data private and giving you full control over the connection. AI engineers can use it to chain models and pipelines into more powerful workflows.
About claude-video-vision
Overview
MCP server that gives Claude Code the ability to watch and understand videos — extracts frames via ffmpeg and processes audio via multiple backends
Links
Topics
claude, claude-code, mcp, mcp-server, video, vision, multimodal, ffmpeg, whisper, gemini
Who Should Use claude-video-vision?
- 1Analyse and describe images as part of your AI workflow
- 2Chain AI models and pipelines through a unified MCP interface
- 3Let Claude orchestrate other AI tools and models
- 4Integrate embeddings, image generation, or speech APIs into your workflow
How to Install claude-video-vision
Before you start
You will need Node.js (v18 or later) installed on your machine — download it from nodejs.org if you haven't already.
- 1Open a terminal (Terminal on Mac, Command Prompt or PowerShell on Windows).
- 2Paste the install command above and press Enter — Node.js will download and run the server automatically.
- 3Add the server to your Claude Desktop config file (see the JSON snippet above) and restart Claude.
The Claude Desktop config snippet above can be copied and pasted directly into your claude_desktop_config.json file — no editing required.