Adding model provenance would be a good idea so you can support multiple. I was wondering if DALLE 2 embeds any metadata in the images but haven't had a chance to investigate.
I also think you should consider adding the option to clarify copyright status ... in particular, allow users to dedicate the image to the public domain via CC0 (https://creativecommons.org/share-your-work/public-domain/cc...), or maybe put it under a CC BY license. May be debatable what the legal status of the images is, but a CC0 dedication would at least get rid of any ambiguity.
I'm not sure if Dalle-2 embeds any metadata, but openAI also has this CLIP model https://openai.com/blog/clip/ that lets you get embeddings. So we will be using this for our search functionality. It works amazingly well. Stay tuned for the search functionality. It is super powerful for prompt engineering.
Thanks for pointing out the copyright status. Most of the images in Discover are crawled on the web, and you can see all the sources under each image. However for the images that people post on our platform, we should definitely add the clarification for copyright status.
I’ve used both DALL-E and stable diffusion extensively, and in my experience DALL-E is more forgiving while stable diffusion is more powerful — with a much higher skill ceiling for writing prompts.
Also, stable diffusion’s open source weights are great for self-hosting which can’t be done with DALL-E.
Is there someone in the illustrator community who is not excited about this meaningful technology? So cool.
You don't have a job anymore. As a start, you will get less money.
Why I have to pay you when GPT 3 can generate prompts for DALL-E?
Send your congratulations to the geniuses which in search of validation and budget optimization are bringing this revolution ahead.
This thing proved for me that the UI design and design in general are the next target.
I have transitioned to pure frontend development, and I am cool for the next three to six years. After this who knows.
On the artistic end. No more digital painting for me.
I will use only analogue mediums, may be someone will pay for a human made picture on the wall....who knows.
Honestly it's not that much different than sampling in music. A few years ago Kayne West tried something similar when instead of paying Aphex Twin for the sample usage of "Avril 14th", he asked John Legend to play basically a carbon copy of the track on a piano[1]. Ultimately he had to pay for the track anyway.
Same here. Did you use my artworks while training your AI model? Cool, pay me licensing rights and residuals. No? Then remove them from your fancy art obfuscator filter.
There's this article "I Went Viral in the Bad Way" from a write from The Atlantic - https://newsletters.theatlantic.com/galaxy-brain/62fc502abcb.... He got criticized so hard for using an AI-generated in his newsletter instead of paying an artist to do the work. People's concerns are valid and real.
Yet, we can also imagine a future that "lower-level" artistic work can be achieved by AI, like blog article thumbnails, while higher-level artistic work will be done by artists, assisted by AI. Artists can get inspiration from AI artwork and reduce some tedious parts of their workflow.
This is such a silly argument. People have been drawing and photoshopping non real depictions of lion elephant hybrids or whatever else for ages. It's already "clogged".
this is great. it's really helpful to see what impact certain keywords and phrases have on the output.
would it be possible for the site to allow toggling between Dalle-2 and the recent Stable Diffusion model release?
edit: I mean, letting you filter between images from Dalle-2 and Stable Diffusion, not actually running the same prompt through the models, which this site obviously does not do
given that the latter has weights available for download, and apparently fits in a consumer GPU, there might soon be more demand for that than for Dalle-2...
This is such a brilliant idea! Most of the images on https://openart.ai/ are from Dalle-2, but we will definitely add Stable Diffusion images thanks to your advice. Thanks!
So much comes from the structure of the prompt. I've played a lot with dalle-mini, because that's all I had access to, and it's interesting to take prompts from this page (OpenArt) and try them over in dalle-mini.
For example, just paste this (From an OpenArt example):
"A young girl with long wavy black hair is riding a motocycle, she is riding through sunset, she has long black boots. Perspective is from back. Pixel art."
The results, in dalle-mini, are so much better than anything I've ever gotten out of it.
And, then, try variation on the theme:
"A young boy with short blonde hair is playing a stratocaster, he is playing in front of a garden stream, he is wearing blue jeans. Perspective is from front. Watercolor."
The expression "OpenArt" sounds like a good fit for true art, as opposed to a form of synthetic visual production of visual arts mimicking art. In a way is hijacking the notion of arts out of humanity.
I wonder how it interprets something like “two lesbians”. There isn’t necessarily anything a picture of might show to indicate that two women are lesbians, unless something intimate is happening. So, I wonder how the program interprets that. Does it just replace lesbian with woman, or is there actually something to indicate that a woman is a lesbian that we might not notice (gaydar?) but a program that has looked at billions of photos has noticed? Does it know us better than we know ourselves?
Beautiful examples... but does anyone else dislike infinite scrolling websites? It feels like a dark pattern intended to maximize time spent on the site.
Has anybody seriously compared Dalle2 with Midjourney? I know that midjourney is a relative no-name but personal testing with the same prompts give me better results (matches expectations) on midjourney.
A quick summary:
Midjourney’s images are usually more aesthetically pleasing than DALL·E, and the model is still adaptable and responsive to stylistic prompts. That makes it a great tool if you want to generate a lot of pleasant images quickly.
Dalle-2 is capable of delivering a wider range of visual styles. As a result, you’re much more likely to craft a surprising or amusing result, or create a never-before-seen image. It’s less likely the image will seem ‘AI-like’. You can also feel more confident that other users haven’t generated very similar images.
[+] [-] akozak|3 years ago|reply
Adding model provenance would be a good idea so you can support multiple. I was wondering if DALLE 2 embeds any metadata in the images but haven't had a chance to investigate.
I also think you should consider adding the option to clarify copyright status ... in particular, allow users to dedicate the image to the public domain via CC0 (https://creativecommons.org/share-your-work/public-domain/cc...), or maybe put it under a CC BY license. May be debatable what the legal status of the images is, but a CC0 dedication would at least get rid of any ambiguity.
[+] [-] cocokechun|3 years ago|reply
Great idea to add model provenance.
I'm not sure if Dalle-2 embeds any metadata, but openAI also has this CLIP model https://openai.com/blog/clip/ that lets you get embeddings. So we will be using this for our search functionality. It works amazingly well. Stay tuned for the search functionality. It is super powerful for prompt engineering.
Thanks for pointing out the copyright status. Most of the images in Discover are crawled on the web, and you can see all the sources under each image. However for the images that people post on our platform, we should definitely add the clarification for copyright status.
[+] [-] GaggiX|3 years ago|reply
[+] [-] Samin100|3 years ago|reply
Also, stable diffusion’s open source weights are great for self-hosting which can’t be done with DALL-E.
[+] [-] O__________O|3 years ago|reply
[+] [-] webmaven|3 years ago|reply
[+] [-] nbzso|3 years ago|reply
You don't have a job anymore. As a start, you will get less money. Why I have to pay you when GPT 3 can generate prompts for DALL-E?
Send your congratulations to the geniuses which in search of validation and budget optimization are bringing this revolution ahead. This thing proved for me that the UI design and design in general are the next target. I have transitioned to pure frontend development, and I am cool for the next three to six years. After this who knows.
On the artistic end. No more digital painting for me. I will use only analogue mediums, may be someone will pay for a human made picture on the wall....who knows.
[+] [-] scottmf|3 years ago|reply
https://twitter.com/RemitNotPaucity/status/15623190045631733...
https://twitter.com/wavefunk_/status/1562468847483244546
https://twitter.com/remi_molettee/status/1562440354955395084
https://twitter.com/pharmapsychotic/status/15625200064741007...
https://twitter.com/TomLikesRobots/status/156243596942050509...
[+] [-] noisymemories|3 years ago|reply
Same here. Did you use my artworks while training your AI model? Cool, pay me licensing rights and residuals. No? Then remove them from your fancy art obfuscator filter.
[1]https://www.theguardian.com/music/2014/aug/26/aphex-twin-kan...
[+] [-] cocokechun|3 years ago|reply
There's this article "I Went Viral in the Bad Way" from a write from The Atlantic - https://newsletters.theatlantic.com/galaxy-brain/62fc502abcb.... He got criticized so hard for using an AI-generated in his newsletter instead of paying an artist to do the work. People's concerns are valid and real.
Yet, we can also imagine a future that "lower-level" artistic work can be achieved by AI, like blog article thumbnails, while higher-level artistic work will be done by artists, assisted by AI. Artists can get inspiration from AI artwork and reduce some tedious parts of their workflow.
[+] [-] hackernewds|3 years ago|reply
Top rate human generated art looks something like so: https://youtu.be/iKBs9l8jS6Q
which the AI generated models will lag for a while still.
[+] [-] woojoo666|3 years ago|reply
[+] [-] userbinator|3 years ago|reply
[+] [-] naillo|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] andy_xor_andrew|3 years ago|reply
would it be possible for the site to allow toggling between Dalle-2 and the recent Stable Diffusion model release?
edit: I mean, letting you filter between images from Dalle-2 and Stable Diffusion, not actually running the same prompt through the models, which this site obviously does not do
given that the latter has weights available for download, and apparently fits in a consumer GPU, there might soon be more demand for that than for Dalle-2...
[+] [-] cocokechun|3 years ago|reply
Any other feedback or suggestions? :)
[+] [-] unknown|3 years ago|reply
[deleted]
[+] [-] pohl|3 years ago|reply
For example, just paste this (From an OpenArt example):
"A young girl with long wavy black hair is riding a motocycle, she is riding through sunset, she has long black boots. Perspective is from back. Pixel art."
The results, in dalle-mini, are so much better than anything I've ever gotten out of it.
And, then, try variation on the theme:
"A young boy with short blonde hair is playing a stratocaster, he is playing in front of a garden stream, he is wearing blue jeans. Perspective is from front. Watercolor."
[+] [-] sebastianconcpt|3 years ago|reply
The expression "OpenArt" sounds like a good fit for true art, as opposed to a form of synthetic visual production of visual arts mimicking art. In a way is hijacking the notion of arts out of humanity.
Any philosophers around to help us with this?
[+] [-] cocokechun|3 years ago|reply
[+] [-] adammarples|3 years ago|reply
[+] [-] irrational|3 years ago|reply
[+] [-] cainxinth|3 years ago|reply
[+] [-] marcodiego|3 years ago|reply
The shadow of the hand is correct.
[+] [-] npteljes|3 years ago|reply
That must mean that the AI copied both the hand and the hand's shadow from the original.
[+] [-] scottmf|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] wkyi|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] tim--|3 years ago|reply
[+] [-] discordance|3 years ago|reply
I'm guessing it refers to a style and I'm curious if there's an index for the different styles.
[+] [-] qiaoshun8888|3 years ago|reply
[+] [-] KaoruAoiShiho|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
A quick summary: Midjourney’s images are usually more aesthetically pleasing than DALL·E, and the model is still adaptable and responsive to stylistic prompts. That makes it a great tool if you want to generate a lot of pleasant images quickly.
Dalle-2 is capable of delivering a wider range of visual styles. As a result, you’re much more likely to craft a surprising or amusing result, or create a never-before-seen image. It’s less likely the image will seem ‘AI-like’. You can also feel more confident that other users haven’t generated very similar images.
[+] [-] andrewstuart|3 years ago|reply
As an aside, has anyone seen any AI generated boardgame boards?
[+] [-] modinfo|3 years ago|reply
https://labs.openai.com/s/Dh2rKLZATnKSt204dbrsAfS5
[+] [-] fragmede|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] NoMoreBro|3 years ago|reply
[+] [-] dukeofdoom|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] kadokaelan|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply
[+] [-] cocokechun|3 years ago|reply