Skip to content

[enhancement]: support lodestones/Chroma #7986

Open
@keturn

Description

@keturn

Is there an existing issue for this?

  • I have searched the existing issues

What should this feature add?

https://huggingface.co/lodestones/Chroma is derived from FLUX.1 [Schnell] with some changes to the model architecture making it about 25% smaller. It retains the Apache license.

Additional Content

Currently the model manager is willing to install a Chroma GGUF, but running it fails like

Missing key(s) in state_dict: "time_in.in_layer.weight", […],
Unexpected key(s) in state_dict: "distilled_guidance_layer.in_proj.bias", […]

Checklist

Based on my experience with making a third-party node for Chroma, here are some things to keep in mind for a full integration:

  • Uses CFG and not FLUX Guidance.
  • Does not use CLIP. (You can pass CLIP to it, because of its history from Schnell, but its not being trained with CLIP and is usable without it.)
  • T5 encoder should not be either padded or cropped to 256/512 tokens…
    • …but it's often useful to put padding tokens on the negative prompt, especially if it's not large on its own.
  • Should be able to use FLUX LoRA, with some tolerance as in feat(LoRA): allow LoRA layer patcher to continue past unknown layers #8059.
  • Should also be able to use Chroma-specific LoRA.
  • torch.compile is much more worthwhile than it was with SDXL. (Then this is likely true of FLUX too? Haven't checked.)
  • Can use FLUX Redux. (But Flex.1 Redux looks pretty bad.)
  • ControlNet - TBD

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions