aszc-dev / ComfyUI-CoreMLSuite

A set of custom nodes for ComfyUI that allow you to use Core ML models in your ComfyUI workflows.
GNU General Public License v3.0
92 stars 8 forks source link

Core ML Suite for ComfyUI

Overview

Welcome! In this repository you'll find a set of custom nodes for ComfyUI that allows you to use Core ML models in your ComfyUI workflows. These models are designed to leverage the Apple Neural Engine (ANE) on Apple Silicon (M1/M2) machines, thereby enhancing your workflows and improving performance.

If you're not sure how to obtain these models, you can download them here or convert your own models using coremltools.

In simple terms, think of Core ML models as a tool that can help your ComfyUI work faster and more efficiently. For instance, during my tests on an M2 Pro 32GB machine, the use of Core ML models sped up the generation of 512x512 images by a factor of approximately 1.5 to 2 times.

Getting Started

To start using custom nodes in your ComfyUI, follow these simple steps:

  1. Clone or download this repository: You can do this directly into the custom_nodes directory of your ComfyUI.
  2. Install the dependencies: You'll need to use a package manager like pip to do this.

That's it! You're now ready to start enhancing your ComfyUI workflows with Core ML models.

Glossary

[!NOTE] Note on Compute Units: For the model to run on the ANE, the model must be converted with the --attention-implementation SPLIT_EINSUM option. Models converted with --attention-implementation ORIGINAL will run on GPU instead of ANE.

Features

These custom nodes come with a host of features, including:

[!NOTE] Please note that using Core ML models can take a bit longer to load initially. For the best experience, I recommend using the compiled models (.mlmodelc files) instead of the .mlpackage files.

[!NOTE]
This repository will continue to be updated with more nodes and features over time.

Installation

Using ComfyUI-Manager

The easiest way to install the custom nodes is to use the ComfyUI-Manager. You can find the installation instructions here. Once you've installed the ComfyUI-Manager, you can install the custom nodes by following these steps:

Manual Installation

  1. Clone this repository into the custom_nodes directory of your ComfyUI. If you're not sure how to do this, you can download the repository as a zip file and extract it into the same directory.
    cd /path/to/comfyui/custom_nodes
    git clone https://github.com/aszc-dev/ComfyUI-CoreMLSuite.git
  2. Next, install the required dependencies using pip or another package manager:

    cd /path/to/comfyui/custom_nodes/ComfyUI-CoreMLSuite
    pip install -r requirements.txt

How to use

Once you've installed the custom nodes, you can start using them in your ComfyUI workflows. To do this, you need to add the nodes to your workflow. You can do this by right-clicking on the workflow canvas and selecting the nodes from the list of available nodes (the nodes are in the Core ML Suite category). You can also double-click the canvas and use the search bar to find the nodes. The list of available nodes is given below.

Available Nodes

Core ML UNet Loader (CoreMLUnetLoader)

CoreMLUnetLoader

This node allows you to load a Core ML UNet model and use it in your ComfyUI workflow. Place the converted .mlpackage or .mlmodelc file in ComfyUI's models/unet directory and use the node to load the model. The output of the node is a coreml_model object that can be used with the Core ML Sampler.

Core ML Sampler (CoreMLSampler)

CoreMLSampler

This node allows you to generate images using a Core ML model. The node takes a Core ML model as input and outputs a latent image similar to the latent image output by the KSampler. This means that you can use the resulting latent as you normally would in your workflow.

Checkpoint Converter

CoreMLConverter

You can use this node to convert any SD1.5 based checkpoint to a Core ML model. The converted model is stored in the models/unet directory and can be used with the Core ML UNet Loader. The conversion parameters are encoded in the node name, so if the model already exists, the node will not convert it again.

[!NOTE] Some models use a custom config .yaml file. If you're using such a model, you'll need to place the config file in the models/configs directory. The config file should be named the same as the checkpoint file. For example, if the checkpoint file is named juggernaut_aftermath.safetensors, the config file should be named juggernaut_aftermath.yaml. The config file will be automatically loaded during conversion.

[!NOTE] For now, the converter relies heavilty on the model name to determine the conversion parameters. This means that if you change the model name, the node will convert the model again. Other than that, if you find the name too long or confusing, you can change it to anything you want.

LoRA Loader

LoRALoader

This node allows you to load LoRAs and bake them into a model. Since this is a workaround (as model weights can't be modified after conversion), there are a few caveats to keep in mind:

LCM Converter

LCMConverter

This node converts SimianLuo/LCM_Dreamshaper_v7 model to Core ML. The converted model is stored in the models/unet directory and can be used with the Core ML UNet Loader. The conversion parameteres are encoded in the node name, so if the model already exists, the node will not convert it again.

[!NOTE] The conversion process can take a while, so please be patient.

[!NOTE] When using the LCM model with Core ML Sampler, please set _samplername to lcm and scheduler to sgm_uniform.

Core ML Adapter (Experimental) (CoreMLModelAdapter)

CoreMLModelAdapter

This node allows you to use a Core ML as a standard ComfyUI model. This is an experimental node and may not work with all models and nodes. Please use with caution and pay attention to the expected inputs of the model.

[!NOTE] While this approach allows you to use Core ML models with many ComfyUI nodes (both standard and custom), the expected inputs of the model will not be checked, which may cause errors. Please make sure to use a model compatible with the expected parameters.

Example Workflows

[!NOTE] The models used are just an example. Feel free to experiment with different models and see what works best for you.

Basic txt2img with Core ML UNet loader

This is a basic txt2img workflow that uses the Core ML UNet loader to load a model. The CLIP and VAE models are loaded using the standard ComfyUI nodes. In the first example, the text encoder (CLIP) and VAE models are loaded separately. In the second example, the text encoder and VAE models are loaded from the checkpoint file. Note that you can use any CLIP or VAE model as long as it's compatible with Stable Diffusion v1.5.

  1. Loading text encoder (CLIP) and VAE models separately
    • This workflow uses CLIP and VAE models available here and here. Once downloaded, place the models in themodels/clip and models/vae directories respectively.
    • The Core ML UNet model is available here. Once downloaded, place the model in the models/unet directory.
      coreml-unet+clip+vae
  2. Loading text encoder (CLIP) and VAE models from checkpoint file
    • This workflow loads the CLIP and VAE models from the checkpoint file available here. Once downloaded, place the model in themodels/checkpoints directory.
    • The Core ML UNet model is available here. Once downloaded, place the model in the models/unet directory.
      coreml-unet+checkpoint

ControlNet with Core ML UNet loader

This workflow uses the Core ML UNet loader to load a Core ML UNet model that supports ControlNet. The ControlNet is being loaded using the standard ComfyUI nodes. Please refer to the basic txt2img workflow for more details on how to load the CLIP and VAE models. The ControlNet model used in this workflow is available here. Once downloaded, place the model in the models/controlnet directory. coreml-unet+controlnet

Checkpoint conversion

This workflow uses the Checkpoint Converter to convert the checkpoint file. See Checkpoint Converter description for more details.

checkpoint-converter

Checkpoint conversion with LoRA

This workflow uses the Checkpoint Converter to convert the checkpoint file with LoRA. See LoRA Loader description to read more about the caveats of using LoRA.

checkpoint-converter+lora

LCM LoRA conversion

Please note that you can use multiple LoRAs with the same model. To do this, you'll need to use multiple LoRA Loaders.

[!IMPORTANT] In this example, the model is passed through the adapter and ModelSamplingDiscrete nodes to a standard ComfyUI's KSampler (not Core ML Sampler). ModelSamplingDiscrete needs to be used to sample models with LCM LoRAs properly.

multiple-loras

Loader with LoRAs

This workflow uses the Core ML UNet Loader to load a model with LoRAs. The CLIP must be loaded separately and passed through the same LoRA nodes as during conversion. See LoRA Loader description to read more about the caveats of using LoRA. Since _loraname and _strengthmodel are baked into the model, it is not necessary to pass them as inputs to the loader.

[!IMPORTANT] In this example, the model is passed through the adapter and ModelSamplingDiscrete nodes to a standard ComfyUI's KSampler (not Core ML Sampler). ModelSamplingDiscrete needs to be used to sample models with LCM LoRAs properly.

loader+lora

LCM conversion with ControlNet

This workflow uses LCM converter to convert SimianLuo/LCM_Dreamshaper_v7 model to Core ML. The converted model can then be used with or without ControlNet to generate images. lcm+controlnet

SDXL Base + Refiner conversion

This is a basic workflow for SDXL. You add LoRAs and ControlNets the same way as in the previous examples. You can also skip the refiner step.

The models used in this workflow are available at the following links:

[!IMPORTANT] SDXL on ANE is not supported. If loading of the model gets stuck, please try using CPU_AND_GPU or CPU_ONLY.
For best results, use ORIGINAL attention implementation.

sdxl

Limitations

[^1]: Unless EnumeratedShapes is used during conversion. Needs more testing.

Support

I'm here to help! If you have any questions or suggestions, don't hesitate to open an issue and I'll do my best to assist you.