A stylized anime model. This model is the result of attempts to get the textures from MouseyMix, a model that I'm very fond of, on characters with more realistic body proportions. The name 'Gyoza' part of 'Store Bought Gyoza' comes from a shortening of "Gyokai" + "Zankuro", two of the artists MouseyMix is trained on. The Store Bought comes from the fact that the 'bones' of this model, i.e. the primary compositional element is SD-Silicon, or to be more specific Silicon-29. SD-Silicon is made using autoMergeBlockWeight, an automated block merger, hence the dumpling is "made by machines". In other words, 工厂做的，商店买的。
The strengths of this model is that you get a very soft illustrative style model with bright texturing but with very powerful scenery and landscapes. V3 and V3.2 are best for generating generating anime girls. VV(5) is the most powerful but it's also the most wild. Use it if you really want some cool and unique stuff.
Everything is on this HuggingFace repository, along with the recipes. I go over every step and include links to the things used if they're publicly available. If you have anything to ask, don't be hesitant to leave a comment. I believe in an open-source approach to Stable Diffusion model mixing so I'll answer any questions you have on the process. It uses Block Weight Merging pretty extensively, so I would recommend you read up on UNet Blocks as a starting point.
But if you want a general rule of thumb, basically what we're doing is splitting the model mixing process into two parts: a picture composition/landscape arrangement component and a textural component. The former we can make by reverse-cosine Block Weight Merging any sort of highly detailed anime model against Silicon-29. I like to use DetailedProject variants. Then, after we tune up the textual component to get the exact sort of texture we want, we cosine Block Weight Merge the textual component against our compositional component. After that, finetune for taste and we're ready.
Additionally, experimentally in 3.5D and culminating in VV, there is a use of a new block weight formula for transferring the anatomy-sense of model and the style/texture-sense of models used during the mid-mix formula. This allows us to do stuff like move the textural oilpaint-like feeling of various CSRB-based mixes onto the body composition of a more pedestrian anime-mix. See this HuggingFace repository for more details and proof-of-concept mixes of some popular models.
Basic universal SD1.5 rules apply. Other than that, these are my personal recommendations:
Any sort of Negative prompt should work more or less fine. I like using the AuroraNegative (and KHFB) because this is a very light but powerful negative-embedding that does not distort the image a lot. However, there are a fair bit of 'bad thing' tokens inherited from the mix components and natural negative prompts should get you where you want in most cases.
Clip skip matters a lot. Clip Skip 2 will generally get you your standard 'novelai' like generation, or the closest thing to it, but with later variations and especially with VV Clip skip 1 will make the checkpoint do some truly dynamic and interesting stuff. I recommend trying out Clip Skip 1 if you're going for a wide aspect angle, maybe it will surprise you.
Version 3.2, 3.5D, and V(5) have baked VAEs. Earlier versions uh... I don't really remember if they do, to be honest. V(5) is recommended be used with its VAE, the lighting effect is extremely strong otherwise and you might get an output that's too bright.
Please don't strip the metadata if you generate with this model, or at least specify that you're using GyozaMix. I don't really care what you do with this model- in any case I couldn't even stop you if I did, but I really enjoy seeing the pretty pictures that people generate with GyozaMix. That's it. Do whatever with this now that I've put it into the world.