TFDAT (Temporal FDAT) is a custom VSR architecture. This is a continuation of my previous VSR architectures, namely TSPANv2. TFDAT is a major step up in every way, providing significantly better quality, temporal coherency, and even inference speed! This arch has support for PyTorch, ONNX, and TensorRT!
This repository only hosts the Pytorch inference code. To train a TFDAT model, you'll want to use traiNNer-redux with the TFDAT config and a video dataset.
For easier inference than the GUI provided in this codebase, try out Vapourkit. To make a video dataset, try my other tool, video destroyer.
-
Clone the repository:
git clone https://github.com/Kim2091/TFDAT
-
Install PyTorch with CUDA: Follow the instructions at pytorch.org.
-
Install required packages:
pip install -r requirements.txt
You can use TFDAT through Vapourkit (preferred), the included GUI (directions below), or the command line.
For high-performance inference, refer to the TensorRT guide.
For an easy-to-use experience with PyTorch or ONNX models, launch the GUI:
python vsr_gui.pyFor more advanced control, you can use the command-line scripts.
Video upscaling (PyTorch):
python test_vsr.py --model_path pretrained_models/tfdat.pth --input path/to/video.mp4 --output path/to/output.mp4Key arguments for test_vsr.py and test_onnx.py:
--video_codec: Specify the video codec (e.g.,libx264,libx265).--crf: Set the Constant Rate Factor for quality (forlibx264/libx265).
Unlike my previous repositories, ONNX conversion is now done within traiNNer-redux instead. Follow these directions: https://trainner-redux.readthedocs.io/en/latest/getting_started.html#convert-models-to-onnx
Thank you to leobby and Bendel for testing the arch!