Yeah, I couldn’t bring myself to fuzz the pictures up too much more. We’ll just ignore the occasional flat screen too. Hah!
I’m using a Flux 1.D variant called New Reality. https://civitai.com/models/161068?modelVersionId=979329
I use gen AI tools to create and post images that I think are beautiful. Sometimes people agree with my selections. 100% of my electricity comes solar power. 🌤️
Yeah, I couldn’t bring myself to fuzz the pictures up too much more. We’ll just ignore the occasional flat screen too. Hah!
I’m using a Flux 1.D variant called New Reality. https://civitai.com/models/161068?modelVersionId=979329
Lookin good! I was just perusing the metadata and got a chuckle out of “wearing a light grey shaved smooth vagina.”
Wildcards and dynamic prompts are a killer feature for getting diversity. You can have ChatGPT output long lists of options and save them as text to pull from, rather than putting the OR options in the prompt. For instance, here’s just a snippet of my 100 line “hair_color_natural_hightlights.txt” wildcard file:
chestnut brown with sun-kissed golden highlights
dark blonde with naturally blended caramel highlights
auburn with subtle coppery undertones and honey highlights
jet black with hints of espresso brown and warm burgundy highlights
chocolate brown with natural honey and toffee highlights
ash brown with delicate pearl-toned highlights
light brown with sunlit blonde highlights
One question though, are you sure you have the dynamic prompts extension enabled in A111? These all look very similar, and the fact that the OR statement is in the prompt metadata still makes me wonder. Typically, the “brunette | ginger | blonde” would get resolved by the dynamic prompt processor before generation, and the image metadata would only show the single selected term by itself.
Lol! Yeah, this workflow idea was one of the main reasons i’m making the transition.
These are very sweet, and love the embroidery details!
All of the SDXL models are too airbrushy plastic to my eye, but you can’t argue with the hand quality over the older 1.5 models. One of the things I’m wanting to try is to start with an XL model to get the basic pose structure in place, then finish with the grittier and more realistic 1.5 models. That introduces a whole new level of adjustment knobs to play with!
Lol! I promise this is a one-time thing. 🤡
Comfy has pros and cons. In A1111 I use a lot of dynamic wildcards that load loras, like “wearing skirt + upskirt_lora OR wearing sweater + downblouse_lora.” That kind of thing is very difficult in Comfy. But the node system gives you a ton of flexibility in other areas. It’s worth fiddling with if you have the time.
Thanks!
Nice!
The method I’ve settled on takes a bit of work to put together. First, I upload PNGs to catbox.moe. This preserves metadata so someone can feed the image into the A1111 PNG Info tab, or by copying the url to https://pngchunk.com.
Next, I upload JPG copies here. That gives me the lemmynsfw hosted url and builds the gallery. Then I put them both together using markdown so that the image gallery is also links to the PNGs. The final format looks like this:
[![](https://lemmynsfw.com/pictrs/image/59c7f6e6-de70-4354-937b-5b82b67fc195.webp)][1]
[![](https://lemmynsfw.com/pictrs/image/88b14211-4464-4cd2-bb28-05e781dd5fc8.webp)][2]
[![](https://lemmynsfw.com/pictrs/image/bf3a69bb-d0f9-4691-b95e-6794880bbc86.webp)][3]
[1]: https://files.catbox.moe/5dsqza.png
[2]: https://files.catbox.moe/dljkxc.png
[3]: https://files.catbox.moe/kcqguv.png
This seems to work well. The only hiccup is that I need to include the first image twice, once in the post body so it shows in the gallery, and once as the post header image. That works okay in the browser, but some lemmy mobile apps show it as a duplicate.
Here’s the final result: https://lemmynsfw.com/post/1372540
In the past, I’ve uploaded to catbox.moe and then provided a link here.
Edit to add that i’m looking forward to seeing this. I haven’t gotten good results with animate diff and realistic models.
Two belly buttons, or one extremely long belly button?
I like the rim lighting on #2
I hope you’re not saying “reverse engineer” like it’s a negative or shady practice. I freely share all of my prompts to help people see what’s working for me, and I like to explore what’s working for everyone else. I’ve had good success with simpler prompts too, like the one for this parrot: https://civitai.com/images/3050333.
No controlnet or inpainting. Everything was generated in one go with a single prompt. I’ll sometimes use regional prompts to set zones for head and torso (usually top 40% is where the head goes, bottom 60% for torso/outfit). But even when I have regional prompting turned off, it will still generate a 3/4 / cowboy shot.
I assume you pulled the prompt out of one of my images? If not, you can feed them into pngchunk.com. Here’s the general format I use with regional prompting:
*scene setting stuff*
ADDCOMM
*head / hair description*
ADDROW
*torso/body/pose*
The loras that are in the top (common) section are weighted pretty low, 0.2 - 0.3, because they get repeated/multiplied in each of the two regional rows. So I think at the end they’re effectively around 0.6 - 0.8.
photo of a young 21yo (Barbadian Barbados dark skin:1.2) woman confident pose, arms folded behind back, poised and assured outside (place cav_rdrguarma:1.1),
(Photograph with film grain, 8K, RAW DSLR photo, f1.2, shallow depth of field, 85mm lens),
masterwork, best quality, soft shadow
(soft light, color grading:0.4)
ADDCOMM
sunset beach with ocean and mountains and cliff ruin in the background ,
(amethyst with violet undertones hair color in a curly layers style:1.2),
perfect eyes, perfect skin, detailed skin
ADDROW
choker ,
(pea green whimsical unicorn print bikini set:1.1) (topless:1.3) cameltoe (undressing, panty pull:1.4)
(flat breast, normal_nipples :1.4),
(tan lines, beauty marks:0.6)
(SkinHairDetail:0.8)
It may be that you’re not describing the clothing / body enough? My outfit prompts are pretty detailed, so I think that goes a long way for Stable Diffusion to determine how to frame things.
Have fun cooking that new GPU!
Wow, this is great!
Here’s the original pixel art from 1987, for the youngus amongus:
Correct. Thats why i post all my images as links to catbox.moe, which doesn’t touch jpg exif or png chunks.
I do post the first image as a jpg on lemmy so that it shows the thumbnail in the thread, but theres a duplicate link to the original png in the post body.
This is a neat idea. I’ll have to look at segmentation nodes for Comfy.