Submitted by IluvBsissa t3_11d9igy in singularity
Cryptizard t1_ja97t9t wrote
Reply to comment by alexiuss in "But what would people do when all jobs get automated ?" Ask the Aristocrats. by IluvBsissa
>By far the best AIs are being controlled by everyone
Midjourney seems like the best text-to-image model by far, way better than stable diffusion. And GPT models are fundamentally too large for regular people to run on their machines. They require expensive, enterprise hardware. There might be a lightweight version eventually, but it will always be inferior to the bigger models that run in the cloud.
alexiuss t1_ja9zh31 wrote
Midjourney is aight for amateurs [because it's really basic use], but otherwise it has fallen insanely FAR, far behind due to SD's controlnet and upscale tools.
Besides weaker toolkit base Midjourney is just a single model, it has an insane amount of censorship, so no self respecting artist who needs to draw human bodies will ever use it. It literally refuses to visualize human butt because it's so stupidly over-censored. You can't teach Midjourney to draw things in YOUR own style as an artist.
behold feet comparison: https://www.reddit.com/r/StableDiffusion/comments/11cpv2x/open_vs_closedsource_ai_art_oneshot_feet/
recent stable diffusion stabilization controlnet breakthrough that does feet and hands: https://www.reddit.com/r/StableDiffusion/comments/11cxy5h/blender_control_net_rig_updated/
recent stable diffusion landscape obliterates any landscape made in MJ: https://www.reddit.com/r/StableDiffusion/comments/11c995v/trees/
Midjourney anatomy, feet, hands are quite mediocre and fail 99% of the time if closeup of the foot or hand needs to be in the image. it's nearly impossible to draw a character holding something in midjourney with closeup of the hand and object. Takes a thousand attempts to get the hand correct-ish in MJ.
Look, at this MJ render of human hands in comparison, the fingers are absolutely fucked:
As for LLMs, we are currently in the stage of "disco diffusion" where we can run small, dreaming LLMs like Pygmalion and Koboldai on google collab with half-decent results.
LLM optimization and fine-tuning is happening right now: https://www.reddit.com/r/singularity/comments/118svv7/what_the_k_less_than_1b_parameter_model
This is very close to a breakthrough we need required to run 10-100 billion param LLMs on personal computers: https://github.com/FMInference/FlexGen
Cryptizard t1_jaa0vrt wrote
I think you are being pretty unfair to MJ. The faces are about a million times better than what stable diffusion can do, which is way more helpful than getting fingers or toes exactly right. It is also not true that it "obliterates any landscape made in MJ".
As far as FlexGen goes, if it is made possible it would take like an hour to process one prompt on a regular GPU. No one would want that.
alexiuss t1_jaa24pk wrote
I'm being super fair as artist who has both.
MJ is aight for base composition concept dev, but running my own sketch through SD produces waaaaaaaaaay better and far more detailed results and none of the fingers/toes turn into fucking potatoes and I can draw people in bikini or revealing or no clothes without getting fucking censored
You're not understanding that MJ runs a double step process, its not a single render.
MJ does 4 images with are low res -> then there is an upscaler running through the image you choose.
The same process is easely replicated in SD where the original render is upscaled with upscaler tool kit [doublestep]. The double or even quadruple upscale > upscape > upscale > upscale path makes far superior, more detailed and more realistic faces in stable diffusion compared to MJ. You can't run the upscale eight times in MJ on a single image, but you can in SD. If you haven't tried to upscale an image in SD eight times you can't tell me faces are better in MJ. There's no way to defeat eight-step upscalers with just a double-step, the 8-step+ produces absolutely superb HD wallpaper art.
Open source demolishes closed source in every situation.
Hardware will catch up soon enough to run LLMS or we'll get better compression tools like flexgen, it's just the beginning LLMS are evolving very fast, the open source LLMS are still being trained. I've tested 6 billion param LLMS and its a bit random compared to GPT3 but it's still quite nice for an uncensored conversation about topics chatgpt refuses to work with.
Cryptizard t1_jaa40i5 wrote
I dunno it seems like you value porn and taboo conversations a lot more than me. Cheers I guess.
alexiuss t1_jaa53pw wrote
No.
Here's the giant problem, in both MJ and OPENAI's GPT3 porn/wrong-think censors are absolute trash, they cause false positives resulting in a very, VERY high % of failure of inquiry even when the topic isn't porn or controversial. If you work with image makers and LLM as much as I do, over 14 hours a day, you would notice a pattern of failure and get incredibly frustrated by it too.
You simply don't notice that you're being censored because you don't pay attention, don't need to work with coherent narrative flow for writing.
MJ censors people in bikini and drawing zombies - the word "corpse" is banned, that is NOT god damn porn. The list of banned words in MJ is huge and they keep expanding it every week with new words without letting anyone know what they are: https://decentralizedcreator.com/list-of-banned-words-in-midjourney-discord/
GPT3 censored concept writing about battles of supervillains vs heroes, which is NOT fucking porn either.
Something doesn't have to be porn for the idiotic, poorly written censor software implemented by corporations to mistakenly assume it's wrong-think. The current censor AIS are absolute, asinine trash. I have specialized scripts that catch the AI output before the result is deleted and its not porn, I assure you. It's just false positives.
You do not want to live in a world where hugs are censored by an AI overlord.
BlueShipman t1_jaacr61 wrote
You shouldn't have to defend uncensored models to these pearl clutching prudes.
As someone who is deep into SD, it demolishes midjourney right now. Controlnet alone...
Cryptizard t1_jaa571j wrote
Fair point.
Spire_Citron t1_jab553p wrote
No way does Midjourney do faces better than SD. I say this as someone who has used both extensively. I've used SD to fix the faces in some of my MJ images. I will say that SD sometimes does pretty bad at smaller/more distant faces in images, but up close it's amazing. You just have to use a decent model with it. Midjourney does have some strength over SD, though. It's very good at creating attractive images just as a default. You could type in absolute nonsense as a prompt and it would probably spit out something pretty.
Viewing a single comment thread. View all comments