SegMoE - The Stable Diffusion Mixture of Experts for Image Generation!

Поділитися
Вставка
  • Опубліковано 2 лют 2024
  • Mixture of experts. Seems hot for AI text generation... but what if you had a mixture of experts for IMAGE generation? Oh. Segmind just did that. Welcome to SegMoE - the mixture of experts for SDXL, SDXL Turbo and Stable Diffusion 1.5.
    Want to support the channel?
    / nerdyrodent
    == Links ==
    huggingface.co/segmind
    github.com/segmind/segmoe
    github.com/segmind/segmoe/blo...
    == More Stable Diffusion Stuff! ==
    * Faster Stable Diffusions with the LCM LoRA - • LCM LoRA = Speedy Stab...
    * How do I create an animated SD avatar? - • Create your own animat...
    * Installing Anaconda for MS Windows Beginners - • Anaconda - Python Inst...
    * Add anything to your AI art in seconds - • 3 Amazing and Fun Upda...
    * Video-to-Video AI using AnimateDiff - • How To Use AnimateDiff...
    * One image Gets You a Consistent Character in ANY pose - • Reposer = Consistent S...
  • Наука та технологія

КОМЕНТАРІ • 43

  • @Mediiiicc
    @Mediiiicc 5 місяців тому +45

    Need one of those experts to specialize in "hands" lol

  • @MarcSpctr
    @MarcSpctr 5 місяців тому +16

    Finally a finetuned model for hands and legs can be used as expert, and maybe some model which can understand stuff like ON, ABOVE, UNDER, INSIDE, etc.

    • @MrGTAmodsgerman
      @MrGTAmodsgerman 5 місяців тому

      What you wanna generate with "inside" in relation to body parts?

  • @paulpardee
    @paulpardee 5 місяців тому +7

    early days, as you say... I don't think this really gives the concept a fair shake. You have models that are better at one thing than others, but all the models currently out today are generalists who just happen to be slightly better at text or prompt adherence, or counting... An expert model would be focused on just text or just counting and those don't exist as far as I know.
    I'd love to see models built for this that have markup built in to tell Moe what they specialize in so it could direct that work to them... It'd be even better if you could have a standard library of models and Moe would dynamically load the best ones based on your prompt.

  • @worthstream
    @worthstream 5 місяців тому +4

    This will be a game changer as soon as it's somewhat optimized. Expecially if they do manage to release a finetuning framework. Using prompts to compute gating functions is an ok starting point, but a (relatively) quick fine tune of that can make the difference.

  • @elihusolano5993
    @elihusolano5993 5 місяців тому +2

    Hope you have a speedy recovery. Thanks for the great content.

  • @ritpop
    @ritpop 5 місяців тому +3

    I don't comment a lot but your content is great. Hope you get better soon.

  • @aimademerich
    @aimademerich 5 місяців тому

    This is phenomenal!!

  • @kariannecrysler640
    @kariannecrysler640 5 місяців тому +3

    So few comments! I’m not used to that lol. Hope you’re good my nerdy friend ✌️💕🤘🥰 🐭

  • @c0nsumption
    @c0nsumption 5 місяців тому

    Fn love that you’re always willing to get dirty when there no community support bud. Thanks for the hard work 🙏🏽

  • @ImAlecPonce
    @ImAlecPonce 5 місяців тому +2

    looks so cool!! I only have 16 gig vram though

  • @stephantual
    @stephantual 5 місяців тому

    Thanks that was fun :) 🤠

  • @sandy66555
    @sandy66555 5 місяців тому +3

    hoping you're feeling back to full rodent normal soon

  • @elihusolano5993
    @elihusolano5993 5 місяців тому +2

    can this new MoE be applied to Loras?

  • @AC-zv3fx
    @AC-zv3fx 5 місяців тому +1

    I wonder if it is possible to create MOE of Pony diffusion, AnimagineXL 3, realistic model and a model that is based on illustrations or traditional paintings.

  • @blacksage81
    @blacksage81 5 місяців тому +2

    I feel like these researchers skipped a whole breakthrough by skipping Qlora, and the myriad of quantization flavors we could have played with and went straight to moe, when nearly all the models are just finetunes of the sd base. Its odd.

  • @fast_harmonic_psychedelic
    @fast_harmonic_psychedelic 5 місяців тому +1

    i guess its a little better but CLIP training with partiprompts would be just as good

  • @yahiiia9269
    @yahiiia9269 5 місяців тому +1

    Could you theoretically use multiple LCM Turbo models?

  • @AC-zv3fx
    @AC-zv3fx 5 місяців тому +1

    I thought those experts must be trained with the model, so it can know what model to choose

  • @nickolaygr3371
    @nickolaygr3371 4 місяці тому

    its like computer processors evolution

  • @fast_harmonic_psychedelic
    @fast_harmonic_psychedelic 5 місяців тому +6

    the whole MOE paradigm seems to me to be theoretically dubious lol

  • @LouisGedo
    @LouisGedo 5 місяців тому +2

    👋

  • @poipoi300
    @poipoi300 4 місяці тому

    Wonder if we could truly consider this MoE. Haven't read the code, but I suspect all this does is amplify bias, probably akin to LCM but instead it's distributed.

  • @sadshed4585
    @sadshed4585 5 місяців тому

    what cuda do you have? my torch is not saying cuda is available

  • @DoorknobHead
    @DoorknobHead 5 місяців тому +2

    ___m_/ o o \_m___
    0:46 Can someone take the Segmoe Ferret to the vet and get that ringworm removed from it's neck? Thanx, in advance.

  • @oquletz
    @oquletz 5 місяців тому

    i don't realy understand what is this. this is a tool to merge models? does it work for sd 1.5?

  • @aimademerich
    @aimademerich 5 місяців тому

    Wow this whole time I thought your voice was AI, get well soon

  • @fast_harmonic_psychedelic
    @fast_harmonic_psychedelic 5 місяців тому +3

    theyre all general models, none of the constituents are experts on any particular thing lol

  • @mattkupka1702
    @mattkupka1702 5 місяців тому

    How was this much different than a checkpoint merge

  • @kallamamran
    @kallamamran 5 місяців тому

    Isn't this just the same as merged models?

  • @renovacio5847
    @renovacio5847 5 місяців тому

    By by Chat GPT4 😂.. i was using it because the image generation.. but know..

  • @JavierGarcia-td8ut
    @JavierGarcia-td8ut 5 місяців тому

    in the SDXL I think you are using too low CFG setting... maybe?

  • @AliasArketer
    @AliasArketer 5 місяців тому +1

    I boggle at what has been done, I boggle at what may yet BE done. We're in territory that we can't show grandparents and convince them it isn't magic anymore than other silly daftards can be convinced it isn't copy-paste.

  • @Guytron95
    @Guytron95 5 місяців тому

    groovy. Too bad they didn't include image-to-image but still groovy.

  • @raymond_luxury_yacht
    @raymond_luxury_yacht 5 місяців тому +2

    24gb humblebrag

  • @erics7004
    @erics7004 5 місяців тому

    Me, with 4gb vram GPU 😢😢

  • @user-oz9tf9zp7k
    @user-oz9tf9zp7k 9 днів тому

    why didn't this take off? Use a face, hands, background etc... expert and get way better images