🔧 Installation#

FastVideo currently only supports Linux and CUDA GPUs. The code is tested on Python 3.10.0 and CUDA 12.4, primarily with NVIDIA H100 GPUs.

Prerequisites#

  • CUDA 12.4 installed and supported

  • Linux operating system

Installation Options#

Option 1: Quick Install#

pip install fastvideo

Option 2: Installation from Source#

1. Install Miniconda (if not already installed)#

wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh
bash Miniconda3-latest-Linux-x86_64.sh
source ~/.bashrc

2. Create and activate a Conda environment for FastVideo#

conda create -n fastvideo python=3.10 -y
conda activate fastvideo

3. Clone the FastVideo repository#

git clone https://github.com/hao-ai-lab/FastVideo.git && cd FastVideo

4. Install FastVideo#

Basic installation:

pip install -e .

Optional Dependencies#

Flash Attention#

pip install flash-attn==2.7.0.post2 --no-build-isolation

Sliding Tile Attention (STA)#

To try Sliding Tile Attention (optional), please follow the instructions in csrc/sliding_tile_attention/README.md to install STA.

Development Environment Setup#

If you’re planning to contribute to FastVideo please see the following page: Contributor Guide

Hardware Requirements#

For Basic Inference#

  • NVIDIA GPU with CUDA support

  • Minimum 20GB VRAM for quantized models (e.g., single RTX 4090)

For Lora Finetuning#

  • 40GB GPU memory each for 2 GPUs with lora

  • 30GB GPU memory each for 2 GPUs with CPU offload and lora

For Full Finetuning/Distillation#

  • Multiple high-memory GPUs recommended (e.g., H100)

Troubleshooting#

If you encounter any issues during installation, please open an issue on our GitHub repository.

You can also join our Slack community for additional support.