.
IM PLEASED TO RELEASE MY NEW CHECK POINT CALLED DREAM DIFFUSION SD3 LIKENESS. ITS ADHERANCE IS SOMETHING IVE NOT SEEN ON ANY SDXL MODEL BEFORE. I GOT A BIT BOARD OF WAITING FOR THIS PROPER SD3 MODEL TO BE REALSED SO .... I WANTED TO DEMONSTRATE WHAT SDXL MODELS CAN PRODUCE WITH SOME EXTRA TRAINING. FEEL FREE TO GO CHECK IT OUT HERE ON CIVITAI, SEE THE IMAGE PROMPTS TO RENDER OUTPUTS AND THE SD3 COMPARASONS IN THE DISCRIPTION. I WOULD LOVE TO HERE YOUR THOUGHTS AND SEE SOME OF YOUR CREATIONS ... THE MODEL CAN BE FOUND HERE... https://civitai.com/models/452950/dream-diffusion-sd3-likeness-by-dice?modelVersionId=504300
Everything You Need To Create Ai Art or video, After Civitai Has Generated Your Images ... Links to Downloads and Files To Install. All the details listed in here are free to use ..... If there is something specific your looking for and I haven't covered it below, pop a message in the comments and I'm pretty sure I'll be able to help you out..
(((HUGE UPDATE BEEN ADDED BELOW FOR THE LATEST AI TOOL)))
You can find all my models to download and use on Civitai https://civitai.com/user/DiceAiDevelopment/models
Stable Diffusion How Ai Runs a 70 Billion Para LLM
My Ai explains in depth where Ai started and predicts the date of singularity.
This is epic as the explainations and demonstrations it gives in the video are very indepth and correct as the year 2024.
A lot of you ask me what PC set up i have to run all the tools at the same time, So thought it would be easier to make a video of me building my rig. I have posted the video below for those who enquire. This was the original build. Since then I have upgraded to 2 GPUs.
I run a Watercooled Aurus Z590 Ultra with 2 RTX 4090 GPUs, with 48GB of Vram. So its possible to run a 70 Billion Para LLM from a home PC. With the differance widely apparant in this video of the jump difference between a 30 B to a 70 B LLM
Whist were on the topic of LLMs, Here is how to hack GPT. Easy to follow instructions. You wont believe how easy it is. I also cover how to protect your GPT to stop yours from being hacked to.
I've had a lot of requests on how to use and add the Turbo samplers that you see me using all the time. So here is a video with the script edit i created to add the 3 Turbo samplers to automatic 11 11
If you want to use my motion module in Animate Diff , your welcome to download it from Civitai. You will see the simple instructions and download here: https://civitai.com/models/317279/animate-any-motion-by-dice
Today I released the new checkpoint version of DREAM DIFFUSION LIGHTNING...Download it here...... https://civitai.com/models/288400/dream-diffusion-xl-and-sd-by-dice
I Have made a mini movie from all free ai tools. Unfortunatly as i produced it in 4k resolution it was to large to upload to Civitai.. So i just loaded it to my youtube. It's turned out really well considering it cost nothing to create. Take a look and see what you think...
For my latest Checkpoint release see my other article : https://civitai.com/articles/4043/the-next-level-checkpoint-is-now-available-on-civitai
Mostly for windows PC users. Also If you want to install a duel system to your PC, I've added the download and ISO for Windows Ubuntu Plasma Edition - (Windows 11 Theme and Tools) https://www.wubuntu.org/
On start up of PC youll have a choice which OS you want to open.
Unlock the full potential of your Stable Diffusion AI with my comprehensive guide on integrating over 260 free art styles. This tutorial is designed for both Forge UI and Automatic 1111 users, ensuring a seamless experience in enhancing your digital art creations. Follow my step-by-step instructions to install these styles effortlessly, and explore the vast possibilities they offer for your projects.
Enhance your digital art with over 260 free art styles for Stable Diffusion, compatible with both Forge UI and Automatic 1111. Our tutorial guides you through the easy installation process and demonstrates how to maximize these styles for your artwork. From a simple 'cat' to complex 'portrait cinematic' and 'pet photography' styles, learn how to apply, edit, and combine styles to refine your creations.
Explore a variety of art categories, including photography, painting, illustrations, 3D renders, and more, each offering unique styles to ignite your creativity. Get tips on using unconventional styles like 'microscopic' and 'experimental' for diverse effects. Discover how to produce high-quality images with 3D render styles and experiment with vector, line art, and design-oriented styles for comprehensive art projects.
Download the new art styles from this link:
https://drive.google.com/file/d/1gtIaIGH2mUFcN2QjIOw7wKfRhQw3PDCv/view
and enhance your digital artwork today.
UDIO MUSIC MAKER
If you havent already checked out Udio.com you deffo should its epic > it writes music to a very high standard and it does it in quick time to. Here are a couple of sample video I put together Udio dont create a video but I just used CapCut to edit my videos to the music.
>>>>>> MY BIG TIP OF 2024 <<<<<<
Now Pinokio is a web-based user interface that allows users to easily install, run and control applications in their web browser. It uses a combination of web technologies and automation to make this process simple and streamlined. With Pinokio, managing applications is easier than ever before, allowing users to focus on the task at hand instead of spending time setting up and configuring their tools. No more messing around with crashing installs with cuda or pytorch issues.
One of the benefits of using Pinokio is its one-click install feature, which makes it easy to set up popular applications like Automatic1111 and ComfyUI with just a single click. This eliminates the need to manually download and configure the applications, saving users time and hassle. With Pinokio, you can quickly and easily get up and running with the tools you need to get your work done.
For example, with Automatic1111 — a popular tool for working with machine learning models — Pinokio's one-click install feature means you can start using the application right away, without having to worry about installing dependencies or configuring the tool yourself. Similarly, with ComfyUI — a user interface for working with Stable Diffusion models — Pinokio makes it easy to get started, so you can focus on your work rather than the setup process
Link for install : https://pinokio.computer/
Youtube Video to check out how great it is:
Stable Diffusion WebUI Forge
Stable Diffusion WebUI Forge is a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference.
The name "Forge" is inspired from "Minecraft Forge". This project is aimed at becoming SD WebUI's Forge.
Compared to original WebUI (for SDXL inference at 1024px), you can expect the below speed-ups:
If you use common GPU like 8GB vram, you can expect to get about 30~45% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 700MB to 1.3GB, the maximum diffusion resolution (that will not OOM) will increase about 2x to 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x to 6x.
If you use less powerful GPU like 6GB vram, you can expect to get about 60~75% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 800MB to 1.5GB, the maximum diffusion resolution (that will not OOM) will increase about 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x.
If you use powerful GPU like 4090 with 24GB vram, you can expect to get about 3~6% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 1GB to 1.4GB, the maximum diffusion resolution (that will not OOM) will increase about 1.6x, and the maximum diffusion batch size (that will not OOM) will increase about 2x.
If you use ControlNet for SDXL, the maximum ControlNet count (that will not OOM) will increase about 2x, the speed with SDXL+ControlNet will speed up about 30~45%.
Another very important change that Forge brings is Unet Patcher. Using Unet Patcher, methods like Self-Attention Guidance, Kohya High Res Fix, FreeU, StyleAlign, Hypertile can all be implemented in about 100 lines of codes.
Thanks to Unet Patcher, many new things are possible now and supported in Forge, including SVD, Z123, masked Ip-adapter, masked controlnet, photomaker, etc.
No need to monkeypatch UNet and conflict other extensions anymore!
Forge also adds a few samplers, including but not limited to DDPM, DDPM Karras, DPM++ 2M Turbo, DPM++ 2M SDE Turbo, LCM Karras, Euler A Turbo, etc. (LCM is already in original webui since 1.7.0).
Finally, Forge promise that we will only do our jobs. Forge will never add unnecessary opinioned changes to the user interface. You are still using 100% Automatic1111 WebUI.
Installing Forge
If you are proficient in Git and you want to install Forge as another branch of SD-WebUI, please see here. In this way, you can reuse all SD checkpoints and all extensions you installed previously in your OG SD-WebUI, but you should know what you are doing.
If you know what you are doing, you can install Forge using same method as SD-WebUI. (Install Git, Python, Git Clone the forge repo https://github.com/lllyasviel/stable-diffusion-webui-forge.git
and then run webui-user.bat).
Or you can just use this one-click installation package (with git and python included).
>>> Click Here to Download One-Click Package<<<
After you download, you uncompress, use update.bat
to update, and use run.bat
to run.
Note that running update.bat
is important, otherwise you may be using a previous version with potential bugs unfixed.
AS of 30/04/2024 these are the latest releases for AI tools.
I have included the links to the usual suspects, Github, Huggingface and few direct websites for the WebUi downloads or the slower online user versions. Some of these listed in this section are epic.. And as usual there all free.....
SCRIPT VERSION 1.5
Face Fusion 2.6.0
Next generation face swapper and enhancer
https://github.com/facefusion/facefusion-pinokio
Hallo
[NVIDIA Only] Hierarchical Audio-Driven Visual Synthesis for Portrait Image Animation
https://github.com/fudan-generative-vision/hallo
Flash Diffusion
Accelerating any conditional diffusion model for few steps image generation
https://gojasper.github.io/flash-diffusion-project/
Chat-With-Mlx
[Mac Onlyl] An all-in-one LLMs Chat UI for Apple Silicon Mac using MLX Framework.
https://github.com/qnguyen3/chat-with-mlx
Phased Consistency Model - generate high quality images with 2 steps
https://huggingface.co/spaces/radames/Phased-Consistency-Model-PCM
Stable Audio
An Open Source Model for Audio Samples and Sound Design
https://github.com/Stability-AI/stable-audio-tools
SillyTavern
AITown
https://github.com/a16z-infra/ai-town
Turn any raw text into a high-quality dataset for AI finetuning
https://github.com/e-p-armstrong/augmentoolkit
LoRA the Explorer
Stable Diffusion LoRA Playground HuggingFace:
https://huggingface.co/spaces/multimodalart/LoraTheExplorer
lavie
Text-to-Video (T2V) generation framework from Vchitect
https://github.com/Vchitect/LaVie
Dust3r
https://dust3r.europe.naverlabs.com/
LlamaFactory
Unify Efficient Fine-Tuning of 100+ LLMs
https://github.com/hiyouga/LLaMA-Factory
Invoke
The Gen AI Platform for Pro Studios
https://github.com/invoke-ai/InvokeAI
Openui
https://github.com/wandb/openui
XTTS
clone voices into different languages by using just a quick 3-second audio clip. (a local version of
https://huggingface.co/spaces/coqui/xtts
RVC
1 Click Installer for Retrieval-based-Voice-Conversion-WebUI
https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI
LCM
Fast Image generator using Latent consistency models
https://replicate.com/blog/run-latent-consistency-model-on-mac
Whisper-WebUI
A Web UI for easy subtitle using whisper model
https://github.com/jhj0517/Whisper-WebUI
Realtime BakLLaVA
llama.cpp with BakLLaVA model describes what does it see
https://github.com/Fuzzy-Search/realtime-bakllava
Realtime StableDiffusion
https://github.com/radames/Real-Time-Latent-Consistency-Model
StreamDiffusion
[NVIDIA ONLY] A Pipeline-Level Solution for Real-Time Interactive Generation
https://github.com/cumulo-autumn/StreamDiffusion
Moore-AnimateAnyone
[NVIDIA GPU ONLY] Unofficial Implementation of Animate Anyone
https://github.com/MooreThreads/Moore-AnimateAnyone
Moore-AnimateAnyone-Mini
[NVIDIA ONLY] Efficient Implementation of Animate Anyone (13G VRAM + 2G model size)
https://github.com/sdbds/Moore-AnimateAnyone-for-windows
PhotoMaker
Customizing Realistic Human Photos via Stacked ID Embedding
https://github.com/TencentARC/PhotoMaker
BRIA RMBG
Background removal model developed by BRIA.AI, trained on a carefully selected dataset and is available as an open-source model for non-commercial use
https://huggingface.co/spaces/briaai/BRIA-RMBG-1.4
Gligen
An intuitive GUI for GLIGEN that uses ComfyUI in the backend
https://github.com/mut-ex/gligen-gui
MeloTTS
High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean
https://github.com/myshell-ai/MeloTTS
Chatbot-Ollama
open source chat UI for Ollama
https://github.com/ivanfioravanti/chatbot-ollama
Differential-diffusion-ui
https://differential-diffusion.github.io/
Supir
ZeST
ZeST: Zero-Shot Material Transfer from a Single Image. Local port of
https://huggingface.co/spaces/fffiloni/ZeST (Project: https://ttchengab.github.io/zest/)
StoryDiffusion Comics
create a story by generating consistent images
https://github.com/HVision-NKU/StoryDiffusion
Lobe Chat
https://github.com/lobehub/lobe-chat
Parler-tts
https://huggingface.co/spaces/parler-tts/parler_tts_mini
Instantstyle
https://huggingface.co/spaces/InstantX/InstantStyle
Openvoice2
Openvoice 2 Web UI - A local web UI for Openvoice2, a multilingual voice cloning TTS
https://x.com/myshell_ai/status/1783161876052066793
SCRIPT VERSION 1.5
IDM-VTON
Improving Diffusion Models for Authentic Virtual Try-on in the Wild
https://huggingface.co/spaces/yisol/IDM-VTON
Devika
https://github.com/stitionai/devika
Open WebUI
User-friendly WebUI for LLMs, supported LLM runners include Ollama and OpenAI-compatible APIs
https://github.com/open-webui/open-webui
CosXL
Edit images with just prompt, an unofficial demo for CosXL and CosXL Edit from Stability AI,
https://huggingface.co/spaces/multimodalart/cosxl
Face-to-all
Diffusers InstantID + ControlNet inspired by face-to-many from fofr (https://x.com/fofrAI) - a localized Version of
https://huggingface.co/spaces/multimodalart/face-to-all
CustomNet
A unified encoder-based framework for object customization in text-to-image diffusion models
https://huggingface.co/spaces/TencentARC/CustomNet
Brushnet
A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion
https://huggingface.co/spaces/TencentARC/BrushNet
Arc2Face
A Foundation Model of Human Faces
https://huggingface.co/spaces/FoivosPar/Arc2Face
TripoSR
https://huggingface.co/spaces/stabilityai/TripoSR
ZETA
Zero-Shot Text-Based Audio Editing Using DDPM Inversion
https://huggingface.co/spaces/hilamanor/audioEditing
Remove-video-bg
https://huggingface.co/spaces/amirgame197/Remove-Video-Background
[NVIDIA GPU ONLY] LGM
LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation
https://huggingface.co/spaces/ashawkey/LGM
vid2pose
Video to Openpose & DWPose (All OS supported)
https://github.com/sdbds/vid2pose
IP-Adapter-FaceID
Enter a face image and transform it to any other image. Demo for the h94/IP-Adapter-FaceID model
https://huggingface.co/spaces/multimodalart/Ip-Adapter-FaceID
Dreamtalk
When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
https://github.com/ali-vilab/dreamtalk
Video2Openpose
Turn any video into Openpose video
https://huggingface.co/spaces/fffiloni/video2openpose2
MagicAnimate Mini
[NVIDIA GPU Only] An optimized version of MagicAnimate
https://github.com/sdbds/magic-animate-for-windows
MagicAnimate
[NVIDIA GPU Only] Temporally Consistent Human Image Animation using Diffusion Model
https://showlab.github.io/magicanimate/
AudioSep
Separate Anything You Describe
(https://huggingface.co/spaces/Audio-AGI/AudioSep)
Tokenflow
Temporally consistent video editing. A local version of
https://huggingface.co/spaces/weizmannscience/tokenflow
ModelScope Image2Video (Nvidia GPU only)
Turn any image into a video! (Web UI created by fffiloni:
https://huggingface.co/spaces/fffiloni/MS-Image2Video)
Text Generation WebUI
A Gradio web UI for Large Language Models
https://github.com/oobabooga/text-generation-webui
MAGNeT
https://github.com/facebookresearch/audiocraft/blob/main/docs/MAGNET.md
VideoCrafter 2
https://github.com/AILab-CVC/VideoCrafter
Bark Voice Cloning
https://huggingface.co/spaces/fffiloni/instant-TTS-Bark-cloning
UPSCALY:
Deforum Automatic 1111 extension:
The video below is running my Dream Diffusion Checkpoint on Automatic 1111
https://civitai.com/models/288400
MY full basics to very advanced tutorial on how to master Deforum
https://civitai.com/articles/5506/deforum-everything-you-need-to-know
Below is a very simple tutorial on how to start out using Deforum to create some amzaing animated renders.
strength : 0: (0.65),25:(0.55)
Translation X
move canvas left/right in pixels per frame
0: (0)
Translation Y
move canvas up/down in pixels per frame
0: (0)
Translation Z
move canvas towards/away from view [speed set by FOV]
0: (0.2),70:(10),340:(15)
Rotation 3D X
tilt canvas up/down in degrees per frame
0:(0), 70:(0), 160:(0.5), 250:(0.5), 300:(0.5)
Rotation 3D Y
pan canvas left/right in degrees per frame
0:(0), 30:(-3.5), 160:(-2.5), 250:(-2.8), 340:(-2), 430:(0)
Rotation 3D Z
roll canvas clockwise/anticlockwise
0:(0), 70:(0.2), 160:(0), 250:(-0.5), 340:(0), 430:(0.5), 520:(0.8)
My Prompt:
{
"0": "Iron man in black armour with blue lighting, vibrant diffraction, highly detailed, intricate, ultra HD, sharp photo, crepuscular rays, in focus",
"70": "Iron man laser projects from hand, surrounded by fractals, epic angle and pose, symmetrical, 3d, depth of field",
"160": "Superman flying in the clouds, fractals, epic angle and pose, symmetrical, 3d, depth of field",
"250": "Spiderman climbing a tall city building, epic angle and pose, symmetrical, 3d, depth of field",
"340": "masterpiece, Batman standing in front of large fire and flame explosions, vibrant colours, Ultra realistic",
"430": "masterpiece, Wolverine has his long claws out from hands, fire burning city, explosions, vibrant colours, Ultra realistic, epic angle and pose, symmetrical, 3d, depth of field",
"520": "masterpiece, Green Goblin flying on his hover board, lightening and black clouds, Ultra realistic",
"710": "masterpiece, The Incredible Hulk with green skin and large muscles lifting a car above his head, ray tracing, vibrant colours, Ultra realistic"
}