cross-posted from: https://lemmy.intai.tech/post/25821

u/Alphyn

The workflow is quite simple. Just load a pic into img2img. Use the same size as the original image, enable the tiles controlnet. Set a high denoise ratio. Run it, maybe feed it back and run it a couple of times more. Then enable ultimate SD upscale, set the ratio to 2x and run it again. Then accidentally run it again. Naturally, you put the result of each run back into img2img and update the picture size. The model is RPGArtistTools3.

  • pablonaj@feddit.de
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    very cool! would be cool if you could add variation to the areas (like a market, port, forts, different architecture, etc) I guess it could be done by combining different images from different prompts.

  • AlexysLovesLexxie@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    wow, that’s actually really cool. Wish I could find some good tutorials on how to use controlnet, especially for poses. Getting real sick of the same 5 poses over and over again, and yet if I try to specify “arm across chest”, my character drops her arms and faces straight towards the camera.

    • karurochari@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I have two ways to do that:

      • I bought few dolls specifically for that. I just take shots with my camera and use them in controlnet.
      • If you lean the basics of blender, there are posable models available which will generate the skeleton, and depth map for hands and feet automatically. The advantage is that those are not reconstructed from the image, but exact, which avoids errors of misclassification.
  • karurochari@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    The final image is actually quite cool! But yes, I think we could use a LoRa to direct the granularity of the details being generated, which is progressively scaled down as we upscale to normalize the generation a bit. This would allow to keep higher denoise ratios and avoiding the “fractal” generation.