Sign In

Everything You Need To Create Ai Art Free links and Installs

Everything You Need To Create Ai Art Free links and Installs

.

IM PLEASED TO RELEASE MY NEW CHECK POINT CALLED DREAM DIFFUSION SD3 LIKENESS. ITS ADHERANCE IS SOMETHING IVE NOT SEEN ON ANY SDXL MODEL BEFORE. I GOT A BIT BOARD OF WAITING FOR THIS PROPER SD3 MODEL TO BE REALSED SO .... I WANTED TO DEMONSTRATE WHAT SDXL MODELS CAN PRODUCE WITH SOME EXTRA TRAINING. FEEL FREE TO GO CHECK IT OUT HERE ON CIVITAI, SEE THE IMAGE PROMPTS TO RENDER OUTPUTS AND THE SD3 COMPARASONS IN THE DISCRIPTION. I WOULD LOVE TO HERE YOUR THOUGHTS AND SEE SOME OF YOUR CREATIONS ... THE MODEL CAN BE FOUND HERE... https://civitai.com/models/452950/dream-diffusion-sd3-likeness-by-dice?modelVersionId=504300

Everything You Need To Create Ai Art or video, After Civitai Has Generated Your Images ... Links to Downloads and Files To Install. All the details listed in here are free to use ..... If there is something specific your looking for and I haven't covered it below, pop a message in the comments and I'm pretty sure I'll be able to help you out..

(((HUGE UPDATE BEEN ADDED BELOW FOR THE LATEST AI TOOL)))

You can find all my models to download and use on Civitai https://civitai.com/user/DiceAiDevelopment/models


Stable Diffusion How Ai Runs a 70 Billion Para LLM

My Ai explains in depth where Ai started and predicts the date of singularity.

This is epic as the explainations and demonstrations it gives in the video are very indepth and correct as the year 2024.

A lot of you ask me what PC set up i have to run all the tools at the same time, So thought it would be easier to make a video of me building my rig. I have posted the video below for those who enquire. This was the original build. Since then I have upgraded to 2 GPUs.

I run a Watercooled Aurus Z590 Ultra with 2 RTX 4090 GPUs, with 48GB of Vram. So its possible to run a 70 Billion Para LLM from a home PC. With the differance widely apparant in this video of the jump difference between a 30 B to a 70 B LLM


If you want to use my motion module in Animate Diff , your welcome to download it from Civitai. You will see the simple instructions and download here: https://civitai.com/models/317279/animate-any-motion-by-dice

Today I released the new checkpoint version of DREAM DIFFUSION LIGHTNING...Download it here...... https://civitai.com/models/288400/dream-diffusion-xl-and-sd-by-dice

I Have made a mini movie from all free ai tools. Unfortunatly as i produced it in 4k resolution it was to large to upload to Civitai.. So i just loaded it to my youtube. It's turned out really well considering it cost nothing to create. Take a look and see what you think...


For my latest Checkpoint release see my other article : https://civitai.com/articles/4043/the-next-level-checkpoint-is-now-available-on-civitai


Mostly for windows PC users. Also If you want to install a duel system to your PC, I've added the download and ISO for Windows Ubuntu Plasma Edition - (Windows 11 Theme and Tools) https://www.wubuntu.org/

On start up of PC youll have a choice which OS you want to open.


Unlock the full potential of your Stable Diffusion AI with my comprehensive guide on integrating over 260 free art styles. This tutorial is designed for both Forge UI and Automatic 1111 users, ensuring a seamless experience in enhancing your digital art creations. Follow my step-by-step instructions to install these styles effortlessly, and explore the vast possibilities they offer for your projects.

Enhance your digital art with over 260 free art styles for Stable Diffusion, compatible with both Forge UI and Automatic 1111. Our tutorial guides you through the easy installation process and demonstrates how to maximize these styles for your artwork. From a simple 'cat' to complex 'portrait cinematic' and 'pet photography' styles, learn how to apply, edit, and combine styles to refine your creations.

Explore a variety of art categories, including photography, painting, illustrations, 3D renders, and more, each offering unique styles to ignite your creativity. Get tips on using unconventional styles like 'microscopic' and 'experimental' for diverse effects. Discover how to produce high-quality images with 3D render styles and experiment with vector, line art, and design-oriented styles for comprehensive art projects.

Download the new art styles from this link:

https://drive.google.com/file/d/1gtIaIGH2mUFcN2QjIOw7wKfRhQw3PDCv/view

and enhance your digital artwork today.

>>>>>> MY BIG TIP OF 2024 <<<<<<

Now Pinokio is a web-based user interface that allows users to easily install, run and control applications in their web browser. It uses a combination of web technologies and automation to make this process simple and streamlined. With Pinokio, managing applications is easier than ever before, allowing users to focus on the task at hand instead of spending time setting up and configuring their tools. No more messing around with crashing installs with cuda or pytorch issues.

Youtube Video to check out how great it is:


Stable Diffusion WebUI Forge

Stable Diffusion WebUI Forge is a platform on top of Stable Diffusion WebUI (based on Gradio) to make development easier, optimize resource management, and speed up inference.

The name "Forge" is inspired from "Minecraft Forge". This project is aimed at becoming SD WebUI's Forge.

Compared to original WebUI (for SDXL inference at 1024px), you can expect the below speed-ups:

  1. If you use common GPU like 8GB vram, you can expect to get about 30~45% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 700MB to 1.3GB, the maximum diffusion resolution (that will not OOM) will increase about 2x to 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x to 6x.

  2. If you use less powerful GPU like 6GB vram, you can expect to get about 60~75% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 800MB to 1.5GB, the maximum diffusion resolution (that will not OOM) will increase about 3x, and the maximum diffusion batch size (that will not OOM) will increase about 4x.

  3. If you use powerful GPU like 4090 with 24GB vram, you can expect to get about 3~6% speed up in inference speed (it/s), the GPU memory peak (in task manager) will drop about 1GB to 1.4GB, the maximum diffusion resolution (that will not OOM) will increase about 1.6x, and the maximum diffusion batch size (that will not OOM) will increase about 2x.

  4. If you use ControlNet for SDXL, the maximum ControlNet count (that will not OOM) will increase about 2x, the speed with SDXL+ControlNet will speed up about 30~45%.

Another very important change that Forge brings is Unet Patcher. Using Unet Patcher, methods like Self-Attention Guidance, Kohya High Res Fix, FreeU, StyleAlign, Hypertile can all be implemented in about 100 lines of codes.

Thanks to Unet Patcher, many new things are possible now and supported in Forge, including SVD, Z123, masked Ip-adapter, masked controlnet, photomaker, etc.

No need to monkeypatch UNet and conflict other extensions anymore!

Forge also adds a few samplers, including but not limited to DDPM, DDPM Karras, DPM++ 2M Turbo, DPM++ 2M SDE Turbo, LCM Karras, Euler A Turbo, etc. (LCM is already in original webui since 1.7.0).

Finally, Forge promise that we will only do our jobs. Forge will never add unnecessary opinioned changes to the user interface. You are still using 100% Automatic1111 WebUI.

Installing Forge

If you are proficient in Git and you want to install Forge as another branch of SD-WebUI, please see here. In this way, you can reuse all SD checkpoints and all extensions you installed previously in your OG SD-WebUI, but you should know what you are doing.

If you know what you are doing, you can install Forge using same method as SD-WebUI. (Install Git, Python, Git Clone the forge repo https://github.com/lllyasviel/stable-diffusion-webui-forge.git and then run webui-user.bat).

Or you can just use this one-click installation package (with git and python included).

>>> Click Here to Download One-Click Package<<<

After you download, you uncompress, use update.bat to update, and use run.bat to run.

Note that running update.bat is important, otherwise you may be using a previous version with potential bugs unfixed.


AS of 30/04/2024 these are the latest releases for AI tools.

SCRIPT VERSION 1.5

AITown

Build and customize your own version of AI town - a virtual town where AI characters live, chat and socialize

https://github.com/a16z-infra/ai-town

image
Augmentoolkit

Turn any raw text into a high-quality dataset for AI finetuning

https://github.com/e-p-armstrong/augmentoolkit

image

LoRA the Explorer

Stable Diffusion LoRA Playground HuggingFace:

https://huggingface.co/spaces/multimodalart/LoraTheExplorer

image

lavie

Text-to-Video (T2V) generation framework from Vchitect

https://github.com/Vchitect/LaVie

image

SCRIPT VERSION 1.3

Dust3r

Geometric 3D Vision Made Easy

https://dust3r.europe.naverlabs.com/

image

SCRIPT VERSION 1.5

LlamaFactory

Unify Efficient Fine-Tuning of 100+ LLMs

https://github.com/hiyouga/LLaMA-Factory

image

SCRIPT VERSION 1.5

Invoke

The Gen AI Platform for Pro Studios

https://github.com/invoke-ai/InvokeAI

image

SCRIPT VERSION 1.5

Openui

Describe UI and see it rendered live. Ask for changes and convert HTML to React, Svelte, Web Components, etc. Like vercel v0, but open source

https://github.com/wandb/openui

image

XTTS

clone voices into different languages by using just a quick 3-second audio clip. (a local version of

https://huggingface.co/spaces/coqui/xtts

image

RVC

1 Click Installer for Retrieval-based-Voice-Conversion-WebUI

https://github.com/RVC-Project/Retrieval-based-Voice-Conversion-WebUI

image

LCM

Fast Image generator using Latent consistency models

https://replicate.com/blog/run-latent-consistency-model-on-mac

image

SCRIPT VERSION 1.3

Whisper-WebUI

A Web UI for easy subtitle using whisper model

https://github.com/jhj0517/Whisper-WebUI

image

Realtime BakLLaVA

llama.cpp with BakLLaVA model describes what does it see

https://github.com/Fuzzy-Search/realtime-bakllava

image

Realtime StableDiffusion

Demo showcasing ~real-time Latent Consistency Model pipeline with Diffusers and a MJPEG stream server

https://github.com/radames/Real-Time-Latent-Consistency-Model

image

SCRIPT VERSION 1

StreamDiffusion

[NVIDIA ONLY] A Pipeline-Level Solution for Real-Time Interactive Generation

https://github.com/cumulo-autumn/StreamDiffusion

image

SCRIPT VERSION 1

Moore-AnimateAnyone

[NVIDIA GPU ONLY] Unofficial Implementation of Animate Anyone

https://github.com/MooreThreads/Moore-AnimateAnyone

image

SCRIPT VERSION 1

Moore-AnimateAnyone-Mini

[NVIDIA ONLY] Efficient Implementation of Animate Anyone (13G VRAM + 2G model size)

https://github.com/sdbds/Moore-AnimateAnyone-for-windows

image

SCRIPT VERSION 1

PhotoMaker

Customizing Realistic Human Photos via Stacked ID Embedding

https://github.com/TencentARC/PhotoMaker

image

SCRIPT VERSION 1.1

BRIA RMBG

Background removal model developed by BRIA.AI, trained on a carefully selected dataset and is available as an open-source model for non-commercial use

https://huggingface.co/spaces/briaai/BRIA-RMBG-1.4

image

SCRIPT VERSION 1.2

Gligen

An intuitive GUI for GLIGEN that uses ComfyUI in the backend

https://github.com/mut-ex/gligen-gui

image

SCRIPT VERSION 1.2

MeloTTS

High-quality multi-lingual text-to-speech library by MyShell.ai. Support English, Spanish, French, Chinese, Japanese and Korean

https://github.com/myshell-ai/MeloTTS

image

SCRIPT VERSION 1.2

Chatbot-Ollama

open source chat UI for Ollama

https://github.com/ivanfioravanti/chatbot-ollama

image

SCRIPT VERSION 1.2

Differential-diffusion-ui

Differential Diffusion modifies an image according to a text prompt, and according to a map that specifies the amount of change in each region

https://differential-diffusion.github.io/

image

SCRIPT VERSION 1.2

Supir

[NVIDIA ONLY] Text-driven, intelligent restoration, blending AI technology with creativity to give every image a brand new life

https://supir.xpixel.group

image

SCRIPT VERSION 1.5

ZeST

ZeST: Zero-Shot Material Transfer from a Single Image. Local port of

https://huggingface.co/spaces/fffiloni/ZeST (Project: https://ttchengab.github.io/zest/)

image

SCRIPT VERSION 1.5

StoryDiffusion Comics

create a story by generating consistent images

https://github.com/HVision-NKU/StoryDiffusion

image

SCRIPT VERSION 1.2

Lobe Chat

An open-source, modern-design ChatGPT/LLMs UI/Framework. Supports speech-synthesis, multi-modal, and extensible (function call) plugin system.

https://github.com/lobehub/lobe-chat

image

SCRIPT VERSION 1.5

Parler-tts

a lightweight text-to-speech (TTS) model that can generate high-quality speech with features that can be controlled using a simple text prompt (e.g. gender, background noise, speaking rate, pitch and reverberation).

https://huggingface.co/spaces/parler-tts/parler_tts_mini

image
SCR
IPT VERSION 1.5

Instantstyle

Upload the picture of an image, and generate images with that image style. Instant generation with no LoRA required

https://huggingface.co/spaces/InstantX/InstantStyle

image

SCRIPT VERSION 1.5

Openvoice2

Openvoice 2 Web UI - A local web UI for Openvoice2, a multilingual voice cloning TTS

https://x.com/myshell_ai/status/1783161876052066793

image

SCRIPT VERSION 1.5
IDM-VTON

Improving Diffusion Models for Authentic Virtual Try-on in the Wild

https://huggingface.co/spaces/yisol/IDM-VTON

image

SCRIPT VERSION 1.5

Devika

Agentic AI Software Engineer

https://github.com/stitionai/devika

imageSCRIPT VERSION 1.2

Open WebUI

User-friendly WebUI for LLMs, supported LLM runners include Ollama and OpenAI-compatible APIs

https://github.com/open-webui/open-webui

image

SCRIPT VERSION 1.5

CosXL

Edit images with just prompt, an unofficial demo for CosXL and CosXL Edit from Stability AI,

https://huggingface.co/spaces/multimodalart/cosxl

image

SCRIPT VERSION 1.5

Face-to-all

Diffusers InstantID + ControlNet inspired by face-to-many from fofr (https://x.com/fofrAI) - a localized Version of

https://huggingface.co/spaces/multimodalart/face-to-all

image

SCRIPT VERSION 1.5

CustomNet

A unified encoder-based framework for object customization in text-to-image diffusion models

https://huggingface.co/spaces/TencentARC/CustomNet

imageSCRIPT VERSION 1.5

Brushnet

A Plug-and-Play Image Inpainting Model with Decomposed Dual-Branch Diffusion

https://huggingface.co/spaces/TencentARC/BrushNet

image

SCRIPT VERSION 1.5

Arc2Face

A Foundation Model of Human Faces

https://huggingface.co/spaces/FoivosPar/Arc2Face

image

SCRIPT VERSION 1.2

TripoSR

a state-of-the-art open-source model for fast feedforward 3D reconstruction from a single image, developed in collaboration between Tripo AI and Stability AI.

https://huggingface.co/spaces/stabilityai/TripoSR

image

SCRIPT VERSION 1.2

ZETA

Zero-Shot Text-Based Audio Editing Using DDPM Inversion

https://huggingface.co/spaces/hilamanor/audioEditing

imageSCRIPT VERSION 1.2

Remove-video-bg

Video background removal tool

https://huggingface.co/spaces/amirgame197/Remove-Video-Background

image

SCRIPT VERSION 1.1

[NVIDIA GPU ONLY] LGM

LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation

https://huggingface.co/spaces/ashawkey/LGM

image

SCRIPT VERSION 1

vid2pose

Video to Openpose & DWPose (All OS supported)

https://github.com/sdbds/vid2pose

image

SCRIPT VERSION 1

IP-Adapter-FaceID

Enter a face image and transform it to any other image. Demo for the h94/IP-Adapter-FaceID model

https://huggingface.co/spaces/multimodalart/Ip-Adapter-FaceID

image

SCRIPT VERSION 1

Dreamtalk

When Expressive Talking Head Generation Meets Diffusion Probabilistic Models

https://github.com/ali-vilab/dreamtalk

image

SCRIPT VERSION 1

Video2Openpose

Turn any video into Openpose video

https://huggingface.co/spaces/fffiloni/video2openpose2

image

MagicAnimate Mini

[NVIDIA GPU Only] An optimized version of MagicAnimate

https://github.com/sdbds/magic-animate-for-windows

image

MagicAnimate

[NVIDIA GPU Only] Temporally Consistent Human Image Animation using Diffusion Model

https://showlab.github.io/magicanimate/

image

AudioSep

Separate Anything You Describe

(https://huggingface.co/spaces/Audio-AGI/AudioSep)

image

Tokenflow

Temporally consistent video editing. A local version of

https://huggingface.co/spaces/weizmannscience/tokenflow

image

ModelScope Image2Video (Nvidia GPU only)

Turn any image into a video! (Web UI created by fffiloni:

https://huggingface.co/spaces/fffiloni/MS-Image2Video)

image

Text Generation WebUI

A Gradio web UI for Large Language Models

https://github.com/oobabooga/text-generation-webui

image

SCRIPT VERSION 1

MAGNeT

MAGNeT is a text-to-music and text-to-sound model capable of generating high-quality audio samples conditioned on text descriptions

https://github.com/facebookresearch/audiocraft/blob/main/docs/MAGNET.md

imageSCRIPT VERSION 1

VideoCrafter 2

[Runs fast on NVIDIA GPUs. Works on M1/M2/M3 Macs but slow] VideoCrafter is an open-source video generation and editing toolbox for crafting video content. It currently includes the Text2Video and Image2Video models

https://github.com/AILab-CVC/VideoCrafter

image

SCRIPT VERSION 1.1

Bark Voice Cloning

Upload a clean 20 seconds WAV file of the vocal persona you want to mimic, type your text-to-speech prompt and hit submit! A local version of

https://huggingface.co/spaces/fffiloni/instant-TTS-Bark-cloning

image------------------------------------------------------------------------------------------------------------

UPSCALY:


Deforum Automatic 1111 extension:

The video below is running my Dream Diffusion Checkpoint on Automatic 1111

https://civitai.com/models/288400

Below is a tutorial on how to start out using Deforum to create some amzaing animated renders.

strength : 0: (0.65),25:(0.55)

Translation X

move canvas left/right in pixels per frame

0: (0)

Translation Y

move canvas up/down in pixels per frame

0: (0)

Translation Z

move canvas towards/away from view [speed set by FOV]

0: (0.2),70:(10),340:(15)

Rotation 3D X

tilt canvas up/down in degrees per frame

0:(0), 70:(0), 160:(0.5), 250:(0.5), 300:(0.5)

Rotation 3D Y

pan canvas left/right in degrees per frame

0:(0), 30:(-3.5), 160:(-2.5), 250:(-2.8), 340:(-2), 430:(0)

Rotation 3D Z

roll canvas clockwise/anticlockwise

0:(0), 70:(0.2), 160:(0), 250:(-0.5), 340:(0), 430:(0.5), 520:(0.8)

My Prompt:

{

"0": "Iron man in black armour with blue lighting, vibrant diffraction, highly detailed, intricate, ultra HD, sharp photo, crepuscular rays, in focus",

"70": "Iron man laser projects from hand, surrounded by fractals, epic angle and pose, symmetrical, 3d, depth of field",

"160": "Superman flying in the clouds, fractals, epic angle and pose, symmetrical, 3d, depth of field",

"250": "Spiderman climbing a tall city building, epic angle and pose, symmetrical, 3d, depth of field",

"340": "masterpiece, Batman standing in front of large fire and flame explosions, vibrant colours, Ultra realistic",

"430": "masterpiece, Wolverine has his long claws out from hands, fire burning city, explosions, vibrant colours, Ultra realistic, epic angle and pose, symmetrical, 3d, depth of field",

"520": "masterpiece, Green Goblin flying on his hover board, lightening and black clouds, Ultra realistic",

"710": "masterpiece, The Incredible Hulk with green skin and large muscles lifting a car above his head, ray tracing, vibrant colours, Ultra realistic"

}

If you enjoy creating video renders then i have a detailed tutorial below on how to to use Deforum in Automatic 1111, to match the render to the beat of your audio.

Download your copy of Dream Diffusion XL and SD

https://civitai.com/models/288400


UPSCALY: Free open source Upscaler which is so simple and perfect outcomes : https://www.upscayl.org/

If you would like my custom made upscale zip files to download, you can see them in my other article on here : https://civitai.com/articles/3900


Flowframes: Fast Video Interpolation for any GPU: https://nmkd.itch.io/flowframes

Below is a video of me giving some examples of how it works.


Lalamu: Add voice to your own video. 30 seconds

Text-to-Speech in German and English

video lip sync with your videos / video templates

https://lalamu.studio/demo/

(UPDATE; Lalamu is now part of CANVA. Shame hey)



Cap Cut:

CapCut Ai Video Editor: Epic install i think its great and it free. lots of Ai tools to fix or create videos: https://www.capcut.com/


ElevenLabs: Free Generative Voice AI: https://elevenlabs.io/


D-ID Creative Reality Studio: Use generative AI to create future-facing videos: https://www.d-id.com/creative-reality-studio/

Below is one of my early attempts to push D-ID to its limits back in 2022


I update this page once a week and add new content that i discover .. Enjoy

39

Comments