š¦ Train Character LoRAs in Illustrious (Step-by-Step Guide)
Hey everyone!
Iāve been asked a lot about how I train my character LoRAs, so I thought Iād finally break it all down for you. This is the exact workflow I use for every LoRA I make ā the same process thatās behind all the characters you see here.
Itās not the only way, but itās whatās been working for me after a lot of trial and error. Hopefully this helps anyone curious about training their own!
š Step 1: Finding the Character
Once Iāve decided who I want to train, itās time to gather images. The more variety, the better.
Hereās where I usually search:
Danbooru (amazing for tagged references and knowing how to correctly tag your characters)
Google (filter by 2MP so you donāt get low quality stuff minimum is 512x512 body & 250xx250 if it's a face/profile picture)
Grabber (an open-source app that searches across multiple art sites)
rule34.xxx / e-hentai / DeviantArt
Original sources like anime, shows, games, 3D models, etc.
š Tip: 16 images is enough to train, but I personally go for 100+. It sounds overboard, but more variety = better results.
āļø Step 2: Prepping the Images
I used to cut out backgrounds and other characters⦠but honestly, itās not necessary.
As long as your dataset isnāt full of repeating junk (text bubbles, duplicate objects, or the same side character multiple times), youāre fine.
Now I only crop when I have to and save myself the headache.
š·ļø Step 3: Tagging in TagGUI
All my images go into one folder, which I open in TagGUI.
Hereās my tagging flow:
Trigger word ā A unique label that will always call the character.
Examples:
4lv1n
,/\lvin
,AlvinC
Important: make sure itās not a normal Danbooru word.
Species ā furry, human, robot, autobot, etc.
Features ā fur/skin color, eyes, scars, unique details.
Clothing ā āpurple collared shirt,ā āwhite shorts,ā ābrown sandals.ā
š” I usually follow Danbooruās tag descriptions for accuracy. If nothing exists, I just make up a consistent one.
ā” Step 4: Auto-Tagging (Huge Time Saver)
Manually tagging everything is painful. Thatās where auto-tagging comes in.
In TagGUI, I use these two:
SmilingWolf/wd-eva02-large-tagger-v3
SmilingWolf/wd-vit-large-tagger-v3
Run them both across all images. After that, I go back and delete any conflicting tags so my custom ones stay clean.
š¦ Step 5: Getting Ready for Training
When tagging is done:
I zip the images + text files together.
Upload the ZIP into Civitaiās online LoRA trainer.
On the trainer page, itāll ask for example prompts. Hereās the type I usually give it:
1boy, solo, trigger word, species, close up, side angle, smirk, looking at viewer
1boy, solo, trigger word, species, clothes, full body
1boy, solo, trigger word, species, naked, big penis, testicles
āļø Step 6: Training Settings (My Defaults)
After lots of testing, these are the numbers I stick to:
Epochs: 20 ā lets me check outputs at 5, 10, 15, and 20.
Num Repeats: Whatever makes total Steps ā„ 1180.
Shuffle Tags: ON
Keep Tokens: 2 ā keeps trigger word + species stable.
Network Dim: 8 ā smaller LoRA file size.
Network Alpha: 4 ā smaller LoRA file size.
Then hit Submit and let it cook.
š Step 7: Reviewing Results
Training usually takes an hour or two. Once itās ready, I download epochs 5-20 normally.
I always compare epochs 5, 10, 15, and 20 side-by-side. One of these will usually be the sweet spot where the LoRA looks both accurate and consistent and it's the one I go with.
š Wrapping It Up
And thatās it! Thatās my full workflow from start to finish.
The most important lessons Iāve learned are:
Gather more images than you think you need.
Make your trigger word unique.
Donāt waste time cutting backgrounds unless you have to.
Always clean up your auto-tags.
Test multiple epochs ā donāt just grab the first one.
This process has saved me so much time and frustration, and I hope it helps you out too. If you try it, let me know how it goes ā or if you discover little tricks of your own along the way!
š§Ŗ How I Avoid āStyle Bake-Inā
A question I get a lot: āHow do you stop the LoRA from learning an artistās style instead of the character?ā Two things make all the difference:
Volume = Neutralization
The more images you train on, the more the model āaverages outā style bias. With 100+ images, you can get an efficient epoch 1 thatās character-true and not have a style baked in.Mix Sources/Styles on Purpose
Blending references from different artists, sites, and mediums (screenshots, fanart, 3D renders, official art, etc.). This forces the LoRA to learn identity features (species, markings, colors, shapes) instead of one stylistic look.
TL;DR: More images + mixed sources = character fidelity without style lock-in.
P.S. I'd just like to clarify, I've stopped using fanart for training and use the original sources.