db0 set up an AI image generator bot both on the Threadverse and Mastodon some time back for anyone to use. All one needs to do is mention it in a comment followed by the text “draw for me” and then prompt text, and it’ll respond with some generated images. For example:

@[email protected] draw for me An engraving of a skunk.

Caused it to reply back to me with:

Here are some images matching your request

Prompt: An engraving of a skunk.

Style: flux

The bot has apparently been active for some time and it looks like few people were aware that it existed or used it — I certainly wasn’t!

I don’t know whether it will work in this community, as this community says that it prohibits most bots from operating here. However, I set up a test thread over here on [email protected] to try it out, where it definitely does work; I was exploring some of how it functions there, and if you’re looking for a test place to try it out, that should work!

It farms out the compute work to various people who are donating time on their GPUs via AI Horde.

The FAQ for the bot is here. For those familiar with local image generation, it supports a number of different models.

The default model is Flux, which is, I think, a good choice — that takes English-like sentences describing a picture, and is pretty easy to use without a lot of time reading documentation.

A few notes:

  • The bot disallows NSFW image generation, and if it detects one, it’ll impose a one-day tempban on its use to try to make it harder for people searching for loopholes to generate them.

  • There appears to me in my brief testing to be some kind of per-user rate limit. db0 says that he does have a rate limit on Mastodon, but wasn’t sure whether he put one on Lemmy, so if you might only be able to generate so many images so quickly.

  • The way one chooses a model is to change the “style” by ending the prompt text with “style: stylename”. Some of these styles entail use of a different model; among other things, it’s got models specializing in furry images; there’s a substantial furry fandom crowd here. There’s a list of supported styles here with sample images.

db0 has encouraged people to use it in that test post and in another thread where we were discussing this, says have fun. I wanted to post here to give it some visibility, since I think that a lot of people, like me, have been unaware that has been available. Especially for people on phones or older computers, doing local AI image generation on GPUs really isn’t an option, and this lets folks who do have GPUs share them with those folks.

  • tisktisk@piefed.social
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    3
    ·
    9 days ago

    Humbly requesting an explanation for the polarized takes on this. How are we so specifically split 50/50 rejoicing for the utility and cursing the ‘slop’ at the same? Someone sway me to a side? I’m addicted to reserving judgement

    • lime!@feddit.nu
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      1
      ·
      9 days ago

      on the one hand, this is an ai horde-based bot. the ai horde is just a bunch of users who are letting you run models on their personal machines, which means this is not “big ai” and doesn’t use up massive amounts of resources. it’s basically the “best” way of running stable diffusion at small to medium scale.

      on the other, this is still using “mainstream” models like flux, which has been trained on copyrighted works without consent and used shitloads of energy to train. unfortunately models trained on only freely available data just can’t compete.

      lemmy is majority anti-ai, but db0 is a big pro-local-ai hub. i don’t think they’re pro-big-ai. so what we’re getting here is a clash between people who feel like any use of ai is immoral due to the inherent infringement and the energy cost, and people who feel like copyright is a broken system anyway and are trying to tackle the energy thing themselves.

      it’s a pretty thorny issue with both sides making valid points, and depending on your background you may very well hold all the viewpoints of both sides at the same time.

      • tisktisk@piefed.social
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        9 days ago

        Both sides having valid points is almost always the case with issues of any complexity. I’m very curious to know why there isn’t a sweeping trump card that ultimately deems one side as significantly more ethical than the other

        Great analysis tho–very thankful for the excellent breakdown unless you used ai to do it or if that ai is ultimately not justifying the means adequately. No actually I’m thankful regardless but I’m still internally conflicted by the unknown

        • lime!@feddit.nu
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          1
          ·
          9 days ago

          no matter your stance on the morality of language models, it’s just plain rude to use a machine to generate text meant for people. i would never do that. if i didn’t take the time to write it, why would you take the time to read it?

          • daniskarma@lemmy.dbzer0.com
            link
            fedilink
            arrow-up
            2
            ·
            8 days ago

            I think there may be two exceptions to that rule.

            1. Accessibility. People who may have issues writing long coherent text due the need to use some different input method (think about tetraplegic people for instance). LLM generated text could be of great aid there.

            2. Translation. I do hate forced translation. But it’s true that for some people it may be needed. And I think LLM translation models have already surpassed other forms of automatic software translation.

            • dil@lemmy.zip
              link
              fedilink
              arrow-up
              1
              ·
              edit-2
              8 days ago

              There are always exceptions/outliers to any rule, it’s basically playing devils advocate to bring them up, I never care for it, like a conversation about someone murdering someone for funsies and saying “but there are cases where people should be murdered, like the joker from batman”

    • tal@lemmy.todayOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 days ago

      Yeah, just one word to refer to all the compatible Fediverse Reddit-alikes.

      Sublinks will be in there too, if they get the ball rolling.

  • LostWanderer@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    4
    ·
    9 days ago

    Good to know, I’ll be sure to block this bot on all of my Fediverse accounts. It’s sad to see more GenAI nonsense.

  • RagnarokOnline@programming.dev
    link
    fedilink
    arrow-up
    3
    arrow-down
    3
    ·
    9 days ago

    This is gold and I’m so thankful for this knowledge.

    I wish I could limit the response to just one picture so I could use this as a meme generator in response to user comments.

    • db0@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      2
      ·
      8 days ago

      I could arrange that somehow as being able to reply with memegens was part of the original idea. But you might still end up with a bad 1shot.

    • tal@lemmy.todayOP
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      edit-2
      9 days ago

      The reason many AI image generators do multiple images is as a simple way to trade compute cycles for quality. The idea is that you generate a couple and pick the best, using your human knowledge of what you intend.

      You could generate it in one place, copy the URL of the best image, and embed it in your response. That’s what I did when I pasted the links to the skunk engraving images in my post; the images were generated elsewhere. I just pasted all four rather than only one, to show what the response looks like.

      The syntax for an inline image on the Threadiverse’s Markdown variant is:

      ![](URL)
      

      I assume that either that syntax or a similar one will work on Mastodon, but I don’t know Mastodon’s syntax, as I don’t use it.

  • amino@lemmy.blahaj.zone
    link
    fedilink
    arrow-up
    3
    arrow-down
    3
    ·
    edit-2
    8 days ago

    with allies like these who even needs Elon Muskies? Grok-ifying the fediverse is sure to lead to a tech utopia /s

  • Krafting@lemmy.world
    link
    fedilink
    arrow-up
    2
    arrow-down
    2
    ·
    9 days ago

    I didn’t know about this, and I tested it on mastodon and it’s awesome! Thanks db0!