sylym / comfy_vid2vid

Apache License 2.0
69 stars 12 forks source link

Vid2vid Node Suite for ComfyUI

A node suite for ComfyUI that allows you to load image sequence and generate new image sequence with different styles or content.

Original repo: https://github.com/sylym/stable-diffusion-vid2vid

Install

Firstly, install comfyui

Then run:

cd ComfyUI/custom_nodes
git clone https://github.com/sylym/comfy_vid2vid
cd comfy_vid2vid

Next, download dependencies:

python -m pip install -r requirements.txt

For ComfyUI portable standalone build:

#You may need to replace "..\..\..\python_embeded\python.exe" depends your python_embeded location
..\..\..\python_embeded\python.exe -m pip install -r requirements.txt

Usage

All nodes are classified under the vid2vid category. For some workflow examples you can check out:

vid2vid workflow examples

Nodes

LoadImageSequence

Local Image

Load image sequence from a folder.

Inputs:

Outputs:

Parameters:


LoadImageMaskSequence

Local Image

Load mask sequence from a folder.

Inputs:

Outputs:

Parameters:


VAEEncodeForInpaintSequence

Local Image

Encode the input image sequence into a latent vector using a Variational Autoencoder (VAE) model. Also add image mask sequence to latent vector.

Inputs:

Outputs:

Parameters:


DdimInversionSequence

Local Image

Generate a specific noise vector by inverting the input latent vector using the Ddim model. Usually used to improve the time consistency of the output image sequence.

Inputs:

Outputs:

Parameters:


SetLatentNoiseSequence

Local Image

Add noise vector to latent vector.

Inputs:

Outputs:

Parameters:


CheckpointLoaderSimpleSequence

Local Image

Load the checkpoint model into UNet3DConditionModel. Usually used to generate a sequence of pictures with time continuity.

Inputs:

Outputs:

Parameters:


LoraLoaderSequence

Local Image

Same function as LoraLoader node, but acts on UNet3DConditionModel. Used after the CheckpointLoaderSimpleSequence node and before the TrainUnetSequence node. The input and output of the model are both of ORIGINAL_MODEL type.


TrainUnetSequence

Local Image

Fine-tune the incoming model using latent vector and context, and convert the model to inference mode.

Inputs:

Outputs:

Parameters:


KSamplerSequence

Local Image

Same function as KSampler node, but added support for noise vector and image mask sequence.


Limits