this post was submitted on 16 Jun 2024
33 points (83.7% liked)

AI Generated Images

7137 readers
290 users here now

Community for AI image generation. Any models are allowed. Creativity is valuable! It is recommended to post the model used for reference, but not a rule.

No explicit violence, gore, or nudity.

This is not a NSFW community although exceptions are sometimes made. Any NSFW posts must be marked as NSFW and may be removed at any moderator's discretion. Any suggestive imagery may be removed at any time.

Refer to https://lemmynsfw.com/ for any NSFW imagery.

No misconduct: Harassment, Abuse or assault, Bullying, Illegal activity, Discrimination, Racism, Trolling, Bigotry.

AI Generated Videos are allowed under the same rules. Photosensitivity warning required for any flashing videos.

To embed images type:

“![](put image url in here)”

Follow all sh.itjust.works rules.


Community Challenge Past Entries

Related communities:

founded 1 year ago
MODERATORS
 

The ComfyUI prompt and workflow is attached to the image: https://files.catbox.moe/s3qufb.png

You can't copy pasta this prompt. There are a few nodes that are specific to SD3 and required.

::: spoiler EDIT: more proof of the chain that lead to this image. They were not all this good. I'm cherry picking for sure and these are just webp's without workflows attached:

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 1 points 4 months ago

Your image doesn't have the workflow for me to have a look. You would need the exact same setup with the unique prompt in all 3 positive positions. I'm not positive my custom sampler setup is fully transferable. If you manually tried to add the prompt, that won't work. It could be how my GPU generates seeds too. I think there is something I've read about that being hardware specific in its determinism.

When I try something new, I lock the seed and build everything incrementally. I was working in a tight space that was not super flexible before it would error, but what I was doing was not anomalous. I picked the best out of a dozen or so variants.

I don't try reproducing other's gens often, and I'm trying to get away from the formulaic prompts by starting with something simple and building out the complexity as I go.

I'm not positive about the mechanism that causes the ComfyUI history to impact subsequent generation images, but I intuitively know there is some kind of mechanism similar to how momentum works inside a LLM. I started off this image session by generating around a dozen images of a doll in grass and left them in the history while I started building the prompt I shared. My big secret for success right now is to first generate pictures of things adjacent to the generation I want to make and then bring them together. I generated a few images of grass, the dolls and a few portraits of women. I don't let any errors sneak into the history unless I really need a variation in one image to jump to where I want to go.

The part I really don't understand is how that momentum is stored in the images I have saved. Like I can drop this post's image into a fresh session and I'll basically start right where I left off, but the history and what I'm calling momentum is not there. However, I have watched hundreds of times where a batched generation builds upon itself incrementally. Like if there are 5 images left and the the present image gets a wonky error, every subsequent image will have an error. My gut feeling is that this must be related to how the clip can be saved and exported, but I could easily be wrong with that one.

With this last session I couldn't get it to do the side view. I'm not on my computer now, but I thought this image was from long before the point when I added the side image attempt or the SD3 logo. There may be an error with those new SD3 modules and how they save workflows too.

Just off the cuff, your image here is weighting the last line of the positive prompt differently. I can see most of the obfuscation techniques that are being used in the pedophilia safety system. There are 'harm, mask, creepy face, and stone eyes" that were layered onto your prompt inside the model itself. You likely also have 'parasitic conjoined twin, and broken bones," but I can't see those directly.

The T5 model is just an LLM, so how that is run in the background will have a big impact, especially how it was quantized. That is a very small model, and I expect it to suck massively. As soon as I saw that I realized I need to track down how to use that model as an LLM to learn its particular idiosyncrasies. A heavily quantized small model like that is going to be a dumbass in so many ways.

Anyways, hope that helps. I'm no expert. I just like exploring deeper than most.