Am I one of the few people who finds these generated pictures really bad? They often have weird and unsettling details when you look closely.
I mean, it’s an incredible achievement in AI that we can generate images at this level, but I don’t want them shown to me on a daily basis while I’m reading blogs.
I sunk ~20 hours and $100 playing with DALL-E since last week and I've had a very different experience. Sure--my first dozen attempts with the engine gave bad results, but once I learned to "speak its language" it got easier to generate highly-polished images. The most realistic results come by appending things like "realistic photograph, 4k, in the style of a fashion magazine" to prompts. I suppose any style would work, as long as the body of source material in that style is (mostly) high-quality.
Here's a couple examples I produced with just a little trial and error. FWIW I have an engineering background and zero design experience.
Maybe they're not perfect, but I'm impressed as hell. Exploring what's possible by wording prompts differently feels very much like using a search engine for the first time. Give it a year. This technology is going places.
I find the images to be incredible, but it's very unsettling when you focus on certain details like hands, feet, and eyes. The hands and feet that it draws are almost always mangled, and while it does a good job of drawing an individual eye, it doesn't seem to draw two eyes in a well coordinated manner, either one eye is bigger than the other, or there is something weirdly unsymmetrical about the eyes that makes the image look creepy.
Something that truly amazes me, is how well Dall-e handles lighting. The lighting and shadowing is really good, comparable to a path tracer, except it isn't really doing the super expensive light simulation that a path tracer does.
I wonder if there’s a potential cottage industry of GANs that then fix up these details — one that knows exactly what a hand should look like and will fix up anything that looks like (or that you identify as) a hand
The article isn't loading for me, so I can't really comment on the images it contains, but I've found telling the ai to apply an impressionistic filter does wonders for removing the unsettling aspect. Obviously that limits you to a specific style of image, but I imagine there are other stylistic filters you might apply that achieve the same goal.
I could spend all day looking at the output of "impressionist cats" and similar queries.
I'm over 1000 credits into Dalle so far (I know, I know..) and you're on the money. You can go a lot further than impressionism, though. Specifying the names of famous illustrators, photographers or artists. Specifying the media used. Lens types. Colorways. Film types. Lighting. The right combinations can yield some incredibly realistic looking things, even faces, and then for the rest of it, there's Photoshop, Gigapixel, and other tools to patch things up. (I've had more luck creating 'elements' with Dalle and then montaging them the old fashioned way.)
The images used in the blog linked by OP are okay but stylistically all over the place. OP acknowledges how difficult good prompts are to write. Beyond that, though, you still need to think like an art director and establish a way to set a common style to avoid jarring the readers, and Dalle alone can't do that.
> Specifying the names of famous illustrators, photographers or artists. Specifying the media used. Lens types. Colorways. Film types. Lighting.
Are training sets prepared with systematic variations in individual axes, as an alternative/addition to tagging each of millions of training images on these axes?
I'm with you. I would hate to see these images all over the place -many are just unpleasant to see.
The cover image generated for the cosmopolitan cover is stunning at first but after seeing it a few times it begins to feel uncomfortable to look at. The uncanny valley is alive and well in many of these images.
The pictures are certainly deep down in the uncanny valley, but I think they would be great for nightmarish games. In fact, game developers (and especially game artists) might be the next profession on the line, to be automated by AI.
I don't understand your assertion. Neither game developers nor artists are in any danger of being automated by AI.
Until Copilot can make the game you want, you cannot replace developers. And until you think AI is ready to replace artists in general, you won't be able to automate game artists.
That's not to say a game with assets largely drawn by AI, and heavily assisted by Copilot, wouldn't be a cool artistic experiment!
assets seem quite possibly ai generatable in the near future. you won't be using AI for the important things yet, but a AA or AAA game has a ton of random assets where for random crap that doesn't matter, but that you need to make the world feel full. that seems like a perfect use of AI assets.
> Am I one of the few people who finds these generated pictures really bad?
Well they're bad at not looking like AI generated art. It's impressive, but I've yet to come across an example that doesn't look like AI generated art. A few seconds of surface level inspection and you can see the weird AI psychodelic circling effect (no idea what the technical name is - eye-ball-ification?)
There's a cyberpunk art Facebook group where some people have taken to sharing AI generated cyberpunk cityscapes, and I've been hard pressed to tell it apart from human art on occasion.
To be fair, I think this is because "cyberpunk cityscape" as an artform has become so cliché and generic, it's easy for an AI to copy it!
The results tend to be residents in the uncanny valley. They are nice if you want something unsettling. They are very impressive, can be very aesthetically pleasing(especially with midjourney) but they look very alien.
Maybe part of the reason we are so impressed with those is because they break our perception of reality. It looks like the renaissance statues that are made from marble but looks like cloth.
Is that fundamentally different from an art director today who hands out assignments to freelancers and then picks out and puts together the best of their output?
It's not about being perfect, it's about having something that doesn't take time to produce. like the article says, searching google and stock image sites looking for a picture that very few people are going to ingest is a huge waste of time.
I would suggest scanning through the r/dalle2 sub-reddit, as the submissions there are rated. There are limitations in the way the current crop of AI generators work, but in the hands of someone who knows these and know what prompts to specify you get completely amazing results that you as a layman can't tell is AI generated (without an expert investigation maybe into pixel-level artifacts).
They are good enough for most people and over time those details will get better until we have no need for illustrators.
Already I see website agencies and bloggers using DALL-E. What I do see is that it is easy to pick out DALL-E generated images, in that its too fantastic. Way over the top to a fault.
it's like the über-modern modern art. the next level of those goofy over-the-top meme images that make the rounds in socials
while you may not like it, you just know that this will be a thing on how to create AI-like images without AI. I used to refer to that as grade school ;P
I mean, it’s an incredible achievement in AI that we can generate images at this level, but I don’t want them shown to me on a daily basis while I’m reading blogs.