I have been struggling with getting a decent image from SD within fewer iterations.

I have played around with different sampling methods, CFG values, and steps. But unable to find a consistent configuration that gives me decent images.

Simple prompts that I am struggling with:

  1. a photo of a puppy, intricately detailed, realistic
  2. drawing of a bowl of fruits, manga style

If I am unable to get good output for simple prompts, I am afraid the output for more complex or abstract prompts will be completely unusable.

Are there any tricks that can reduce the iterations to give decent images? Any guidance would be really appreciated.

Thanks!

  • Windex007@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    1 year ago

    The really dynamite images you see on social media are generally not generated with the default models.

    If you’ve messed with the iterations and cfg and still aren’t getting anything close to what you want, I’d try tweaking prompts and for sure I’d be trying the same prompt across multiple models.

    • AusatKeyboardPremi@lemmy.worldOP
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Okay. I did not play with anything beyond the default models. Any suggestions for non-default models?

      As for tweaking prompts, yes, I am already doing it. But I still made the post to be sure I am not missing anything.

      Thanks for your answer and suggestions. :-)

      • Windex007@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        I’d check out https://civitai.com/ and see if any seem to align with what your intentions are.

        Basically, people will augment the training of existing models through various means.

        One of your prompts said something about anime or manga and I am aware that there are many models trained specifically to be good at that.

            • AusatKeyboardPremi@lemmy.worldOP
              link
              fedilink
              arrow-up
              2
              ·
              1 year ago

              I actually was able to generate much better images and learn about LoRA, hypernetworks, etc. thanks to the website you shared. :-)

              Having said that, Stable Diffusion is a bit too cumbersome and tedious when compared to MidJourney. But it is FOSS, and easier to get started with thanks to tools from Automatic1111.

              Looking forward to SDXL, hopefully it alleviates some of the pain points.

              • Windex007@lemmy.world
                link
                fedilink
                arrow-up
                2
                ·
                1 year ago

                I haven’t tried midjourney so I don’t really have a frame of reference. After some practice and getting a good flow going, I was surprised how quickly I could get things to a point that I liked them. Does midjourney do in painting? I’ve been using SD for creating game assets. I’m a trash artist but if I scribble a shit version and img2img with a batch size of like 20… I almost always get something REALLY close to what I need. My usecases are super dependent on inpainting so it’s kinda a must have.

                There are a few other tools that leverage SD. If you’re curious look into retro diffusion for aesprite, super neat workflow there

                • AusatKeyboardPremi@lemmy.worldOP
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  1 year ago

                  I don’t think MJ does inpainting yet, or at least in an accessible way like SD.

                  I haven’t used Aseprite but Retro Diffusion looks really cool and useful.

                  I was initially trying to generate retro/pixel art with the help of prompts, but it was mostly hit or miss. I then found a few webui extensions, like sd-webui-pixelart, that got me closer to the goal.