Sign In

T5xxl Google FLAN from FP32 (CLIP ONLY)

20
220
5
Updated: Nov 20, 2024
base model
Verified:
SafeTensor
Type
Checkpoint Trained
Stats
99
0
Reviews
Published
Oct 28, 2024
Base Model
Other
Training
Steps: 10,000,000
Epochs: 10,000
Hash
AutoV2
AFB1671887
SDXL Training Contest Participant
Felldude's Avatar
Felldude
License:

T5xxl Google FLAN from FP32

NOTE: It appears as when the decoder blocks are pruned from the model it looses the ability to interact with FLUX or SD 3.5 - When I merged the models with the full 42GB file this did not happen.

Trimmed Sizes:

  1. FP32 18GB (42GB with Decoder Weights)

  2. FP16/BF16 9GB

  3. FP8 4.5GB (I do not recommend the FP8 version it seems to have lost to much precision)

I created a tool to extract the FLAN T5.

  • I have a seed to seed image comparison on FLAN T5

  • Speed increase may be from reduced VRAM Load and or reduced file size decreasing time to load in the case of CPU offloading.

  • All Models built of the FULL FP32 FLAN model

  • I have had issues with GGUF quantization using either the T5 header or T5 decoder header, and have been unable to quantize the model using GGUF

Per the Apache 2.0 license FLAN is attributed to Google