For instance if I am adorning a random blog post, I’d rather get a free stock image in the header than pay an artist for a new professional photo, as I don’t think my readers care that much (see “I replaced all our blog thumbnails using DALL♾ 2” for an example).Įspecially if this site was monetized and the big picture was just there to make you scroll further down and get more engagement and ad views.įor cases where the artist’s vision matters, like original paintings for decorating my home, or the panels for a graphic novel, I think StableDiffusion or DALL-E 2 for that matter are far away from beating humans. However, for those cases I think we already had stock images. After lots of discussion in Reddit and at parties, I will try to summarize my current opinion on that topic.įor use-cases where having a human artist brings the least value, I think text-to-image models will dominate the market. Given this context, many people are concerned some artists may lose their jobs. I don’t see any obvious blockers or barriers to the next generation of models being even bigger or understanding style better. This would also apply for frames for an animated movie or a storyboard.Īre we really that close to something so big? I feel like the technology is there if enough compute and budget were allocated, but I am not sure whether someone will do it. I can only imagine what applications artists and other users will come up with in the near future by leveraging StableDiffusion’s embeddings and its text-to-image capabilities, let alone whatever the next generation of models will be able to do.Įxtrapolating from how much this field has grown in the last 18 months, I wouldn’t be surprised if in 2 more years you can write a script for a comic book, feed it to some large language encoder and a text-to-image model like this, and get a fully illustrated, style-coherent graphic novel. I am tired of repeating the same old speech, but thinking back to how primitive models were just a year and a half ago with DALL-E and other VQVAE, this is completely insane. Since it is open source and anyone who has 5GB of GPU VRAM can download it (and Emad Mostaque, Stability.ai’s founder has come out and said more efficient models are coming) to get unlimited uses, expect to keep seeing headlines about AI art for a while. It is similarly powerful to DALL-E 2, but open source, and open to the public through Dream Studio, where anyone gets 50 free uses just by signing up with an email address. What a week, huh? A few days ago, Stability.ai released the new AI art model Stable Diffusion.
0 Comments
Leave a Reply. |