System requirements wise, for a while I used it on a 1050Ti with 4GBs of VRAM. I wouldn't recommend going any lower than that. An RX580 with 8GB VRAM does wonders at a similar secondhand price point (if there isn't any crypto hype going around where you are)
Civitai has a really good selection of models, loras, and other resources
Models
Models are basically the brains of Stable Diffusion. They are the data SD uses to learn what your prompts mean.
The built-in models that come with Stable Diffusion are really bad for porn. Don't use them. In fact don't use them at all unless you're training your own models, there are better SFW models.
Here are some of my personal favourites:
Anime
MeinaHentai is a great model to start with. Compared to other models it's really easy to prompt
AOM3 also does really well, though it might be a little more difficult to guide
Berry Mix (Pre-mixed version here) can also work pretty well, depending on what you want to do. AFAIK it uses rule34 tags instead of danbooru, so it probably won't work all too well with prompts used for the above ones
Realistic
Uber Realistic Porn Merge is the only realistic model I know of that does hardcore stuff. It's unfortunate problem is that it's REALLY DAMN HARD TO USE
VAEs
VAEs are mostly used for finetuning colors, sharpness, what have you. Some models come with a VAE builtin, but for ones that don't, it's recommended to have one on hand.
"Anything VAE", "Orangemix VAE", and "NAI Leak VAE" are the same exact thing under different names. If you already have one on hand, don't bother with the others. Most VAEs are renamed versions or modifications of this one.
Waifu Diffusion's kl-f8-anime2 is also a pretty good one. It doesn't require Waifu Diffusion.
LoRAs teach models about concepts (characters, clothing, environments, style, ...) they might not know about. There are a LOT of them, so feel free to browse Civitai to find ones you might want.
LoRAs tend to be specific for families of models, or at the very least styles (using anime LoRAs on realistic models tend to be a bad idea), but there are a fair few that will work across the board.
Locon and LyCORIS are newer formats of LoRAs. Not sure on the technical differences between them, but they will not work out of the box and need an extension such as https://github.com/KohakuBlueleaf/a1111-sd-webui-lycoris to get working
Textual Inversions / Embeddings and Hypernetworks
These are mostly obsoleted by LoRAs. There are a few embeddings such as Deep Negative and EasyNegative that are still quite useful, but in most cases you'll want to use LoRAs instead.
I find Automatic1111's WebUI to be pretty confusing and difficult to work with. I've switched to Easy Diffusion and I love it waaay more. It's also a 1-click install program, all dependencies are already included
Are these softwares all free to download/use?
Also, how does one start doing this? do i just need the WebUi or do i need extra files to feed into it and stuff?
For the how: Aside from the webui (or whatever else you're running the stable diffusion code on), you'll need models that instruct it what to create. The builtin models are terrible for porn.
I've been meaning to figure this out. So the commonly used ones are labeled WebUI, but just how much of the content goes to the web itself?
If I wanted to train on images that I don't want going online, will they? Or will the products that I create end up online, or does all this stay local?
By Web UI it means that the graphical part of it -- where you write your prompt and hit generate -- is running inside your browser and not as a separate window or command line. Everything is kept on your own computer unless you explicitly tell it to open up remote access.
It was probably a dumb question, but I wanted to clarify just the same, thanks. When ya get these kinds of pics under strict confidence, you don't wanna risk breaking that in any way.
Is it just me or do all the AI generated porn images have something in common that makes them immediately recognizable? I am not quite sure what exactly it is but they do all have some quality in common.
That might be part of it but I think there is also a sort of sameness, certain aspects that are always the same in the generated images and more varied in real ones. Possibly related to the averaging of all the input data.
This is a great guide and was really helpful when I decided to experiment to see how this works.
A couple of things that confused me when trying this out that might save the next person some time:
where to put models etc
*.ckpt and *.safetensors files live in stanle-diffusion-webui/models/stable-diffudion
These will automatically be loaded when you new start wrbui-usr.bat
how to change models
This took me waaay longer to figure out than I'd like to admit. There's a drop down top left of the webui to select the model after you restart
I find the range of models, loras, checkpoints, extensions etc overwhelming. Im still not sure exactly what each of these do and which ones I'd need. Eg: Whats a checkpoint for?
prompt writing is clearly a fine art and can drive you mad.
For both 3&4 I found civitai.com/images to be a fantastic resource. Browse through the images for styles or images you like and most of them will have the resources used and generation data there to recreate it. I found this to be a great starting point, particularly for negative prompts.
deformity
Deformed faces have mostly gone away for me by changing this webui setting: settings> face restoration > code former weight = 0
Just need to figure out hands and phantom limbs now...
Forgot to add, when experimenting with prompts. Use a fixed seed number so you can see how your prompt changes effect the image between each generation
I remember a quick and uncomplicated site that gave access to some models and took a prompt. It wasn't used too much and pretty fast if you selected a non default model. Sadly, I don't remember the name but I miss it. It didn't have a whole lot of bloated J's framework in the frontend like the ones mentioned above.
I use 600x800 (or 800x600) and 50 steps when experimenting with prompts. Then when i get one i like, I lock the seed, maybe do a few .01-.03 variations. Then lock the variation seed and turn on 2x HighRes Fix. That outputs a 1200x1600 that very closely looks like what I expected. (I'm using a 3090 with 24GB vram)
In my experiments, i found that doing quick and smaller tests with a low number of steps, then increasing it for the highres would change the output too much. I settled on this so that theres less options to toggle back and forth.
ROCm is flaky in regular consumer GPUs at the best of times. I'm surprised you could even get that far on a steam deck.
Try the command line arg --opt-sdp-attention. You might also want to try out --medvram or --lowvram (4GBs is considered low when it comes to AI). although I have a feeling it's just because of the custom nature of the deck's APU.
Your best bet would be to search for builds of ROCm, PyTorch and Torchvision that are specifically made for the deck, if such things even exist.
Is anyone here using SD with Blender or for stuff other than NSFW too? Looking to see if a laptop embedded RTX 4050 or 4060 is viable? What is the real limit here? Like can anyone tell me something like "8gbv just can't do (XYZ)" or the iteration time becomes so long the workflow is not practical/sw crashes. Anyone running Linux on a 4050/4060? What kind of impact is there with storage speeds, what size of DDRx is used in practice?
I tried to install sd webui,
Installed python, git.
When I run the run.bat
Error keeps on coming
AttributeError: module 'OS' has no attribute 'statvfs'
Tried re installation of python with environment variable and max limit disable, still run into same attribute error.
I am on windows 11, with nvidia
Any pointers on what I am doing wrong
I recently came across Krita's AI plugin. It's pretty easy to install (especially if you have nVidia or ComfyUI already set up). And lets you easily fine tune your images.
Anyone else use the OpenPose control net? I'm a little confused on the head portion and which control node is doing what. Every thing else about the editor makes sense and is intuitive. Do the outer most nodes on the head affect head tilt and rotation? And are the inner nodes eye placement? this seems to be the case but I'd rather not end up confused and mad when I'm inevitably wrong and their heads are backwards and eldritch abominations. lol
This is a question which bothers me about any AI image generation, but here people are focused on doing humans so I feel it's a good place to ask as with humans it hits harder:
Let's assume I download a recommended setup and I'm a total noob sitting down to generate from a prompt.
How many misshapen, badly generated disturbing chaos mutants from beyond reality do I need to see before the AIs return a somewhat satisfying result? Is an "unsuccessful image" just a person with a blurred face or somewhat off fingers, or are we talking about full-blown body horror?
Since no one has answered yet I'll chime in with my experience using the Stablediffusion thing with all the 111's in it a few months ago.
It was super easy to use. I would set a prompt like "25 year old woman nude at the beach. Blonde hair blue eyes, thin small perky breasts, shaved pubic area" and run usually 10 batches of 5 images each. It would take 5-10 minutes usually, I have a 1080 TI, and I'd probably get 5-10 ""good" images and the rest would be trash. Some would have weird extra arms or be snake people with weird torsos. The most common issue I would have would be nipple and vulva placement, it can be weird sometimes. Not uncommon to have extra or no nipples, or extra breasts. Lots of barbie type pubic areas would show up.
I think part of my issue might have been the HD upscaling I was using, as I would see a quick glance of the initial render but by the time it upscaled it went funky.
I honestly just chocked it up to my lack of technical knowledge and possibly bad prompt writing.
However I do feel it was incredibly easy to generate at least some decent stuff for someone that has no coding experience or anything.