Spaces:
Build error
WhisperFusion
Seamless conversations with AI (with ultra-low latency)
Welcome to WhisperFusion. WhisperFusion builds upon the capabilities of the WhisperLive and WhisperSpeech by integrating Mistral, a Large Language Model (LLM), on top of the real-time speech-to-text pipeline. Both LLM and Whisper are optimized to run efficiently as TensorRT engines, maximizing performance and real-time processing capabilities. While WhiperSpeech is optimized with torch.compile.
Features
Real-Time Speech-to-Text: Utilizes OpenAI WhisperLive to convert spoken language into text in real-time.
Large Language Model Integration: Adds Mistral, a Large Language Model, to enhance the understanding and context of the transcribed text.
TensorRT Optimization: Both LLM and Whisper are optimized to run as TensorRT engines, ensuring high-performance and low-latency processing.
torch.compile: WhisperSpeech uses torch.compile to speed up inference which makes PyTorch code run faster by JIT-compiling PyTorch code into optimized kernels.
Getting Started
- We provide a pre-built TensorRT-LLM docker container that has both whisper and phi converted to TensorRT engines and WhisperSpeech model is pre-downloaded to quickly start interacting with WhisperFusion.
docker run --gpus all --shm-size 64G -p 6006:6006 -p 8888:8888 -it ghcr.io/collabora/whisperfusion:latest
- Start Web GUI
cd examples/chatbot/html
python -m http.server
Build Docker Image
- We provide the docker image for cuda-architecures 89 and 90. If you have a GPU with a different cuda architecture. For e.g. to build for RTX 3090 with cuda- architecture 86
bash build.sh 86-real
This should build the ghcr.io/collabora/whisperfusion:latest
for RTX 3090.
Contact Us
For questions or issues, please open an issue. Contact us at: [email protected], [email protected], [email protected]