Sign In

T5xxl Google FLAN from FP32 (CLIP ONLY)

20

220

5

Updated: Nov 20, 2024

base model

Verified:

SafeTensor

Type

Checkpoint Trained

Stats

99

0

Reviews

Published

Oct 28, 2024

Base Model

Other

Training

Steps: 10,000,000
Epochs: 10,000

Hash

AutoV2
AFB1671887
SDXL Training Contest Participant
Felldude's Avatar

Felldude

License:

T5xxl Google FLAN from FP32

NOTE: It appears as when the decoder blocks are pruned from the model it looses the ability to interact with FLUX or SD 3.5 - When I merged the models with the full 42GB file this did not happen.

Trimmed Sizes:

  1. FP32 18GB (42GB with Decoder Weights)

  2. FP16/BF16 9GB

  3. FP8 4.5GB (I do not recommend the FP8 version it seems to have lost to much precision)

I created a tool to extract the FLAN T5.

  • I have a seed to seed image comparison on FLAN T5

  • Speed increase may be from reduced VRAM Load and or reduced file size decreasing time to load in the case of CPU offloading.

  • All Models built of the FULL FP32 FLAN model

  • I have had issues with GGUF quantization using either the T5 header or T5 decoder header, and have been unable to quantize the model using GGUF

Per the Apache 2.0 license FLAN is attributed to Google