High-performance speech recognition MCP server based on Faster Whisper, providing efficient audio transcription.
What is Fast-Whisper-MCP-Server?
Fast-Whisper-MCP-Server is a high-performance speech recognition server based on Faster Whisper, designed to provide efficient audio transcription.
How to use Fast-Whisper-MCP-Server?
To use the server, clone the repository, install the required dependencies, and start the server using the provided script. After that, you can configure it using a compatible application such as Claude Desktop.
What are the key features of Fast-Whisper-MCP-Server?
- Integration with Faster Whisper for efficient speech recognition
- Batch processing acceleration for faster transcription
- Automatically enable CUDA acceleration if available
- Support for multiple model sizes (from miniature to large - v3)
- Output formats include VTT subtitles, SRT and JSON.
- Model Instance Caching to Avoid Duplicate Loads
- Dynamically adjusts batch size based on GPU memory
What are the application scenarios of Fast-Whisper-MCP-Server?
- Transcribe audio files for content creation
- Real-time speech recognition in applications
- Batch processing and analysis of multiple audio files
Fast-Whisper-MCP-Server FAQ?
- What are the system requirements?
Requires Python 3.10+, Faster Whisper, and PyTorch with CUDA support for optimal performance.
- Can it handle multiple audio files at the same time?
Yes! It supports batch transcription of audio files in folders.
- Is there a GUI?
It is currently command line based, but can be integrated into GUI applications like Claude Desktop.
- ¥Download for freeDownload after commentDownload after login
- {{attr.name}}: