Тёмный

That man singing doesn't exist 

2kliksphilip
Подписаться 650 тыс.
Просмотров 309 тыс.
50% 1

DALL-E Mini, Midjourney, DALL-E 2. I pit these 3 AI image generators against each other to see which reigns supreme. Spoiler alert: DALL-E 2 wins.
Try DALL-E Mini here: www.craiyon.com/
0:00 - 3 tools tested
2:20 - Stylised
4:54 - Realistic
5:49 - Martian base
7:16 - Text
8:12 - The vertical mattress challenge
9:27 - Kitchens
10:22 - Cities
11:30 - Dog sniffing lamp post
12:14 - Variants on existing images
12:42 - Upscaling limitations
14:03 - The future

Игры

Опубликовано:

 

23 май 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 805   
@asdasdasd9320
@asdasdasd9320 Год назад
Philip has such an abstract idea of a bed, that even DALL-E 2 can't handle it
@BonJoviBeatlesLedZep
@BonJoviBeatlesLedZep Год назад
I'm still in doubt that that's his actual bedroom and he comfortably sleeps on that. He lives with his girlfriend right? He must surely be sleeping in an actual sane bed with her most nights.
@YTnamesAreBS
@YTnamesAreBS Год назад
This comment made me exhale quickly from my nose, which is the highest honor I can bestow on you.
@Blox117
@Blox117 Год назад
i tried inputting "honest woman" but dalle 2 told me to ask for something reasonable
@iR0CKt
@iR0CKt Год назад
@@Blox117 Need to be stylized or something :D
@ChrisD__
@ChrisD__ Год назад
@@iR0CKt That's right, you can only get yourself a furry or anime GF.
@cortster12
@cortster12 Год назад
You know the crazy part? We're not reaching the endgame, these tools are still in their infancy. We're like people looking at room sized computers in the 1950s and being amazed. Because it was amazing, the future just ended up being unfathomable.
@vadiks20032
@vadiks20032 Год назад
aren't they written in python imagine the speed if they were written on C languages
@fueledbycoffee583
@fueledbycoffee583 Год назад
@@vadiks20032 written in python calling c++ code. So would be about the same.
@4rumani
@4rumani Год назад
this is the end lol, the ai boom is over and we're headed towards a long long ai winter
@cortster12
@cortster12 Год назад
@@4rumani I will remember this comment when it's completely wrong.
@BlackParade01
@BlackParade01 3 месяца назад
​@@4rumanioh boy how wrong you were
@vankata69exe45
@vankata69exe45 Год назад
philip is an ai with very good text to speech at this point
@supremesurvivor
@supremesurvivor Год назад
This is certainly one of my favorite videos on youtube, but it's so scary, unnerving, that I can't even describe what I'm feeling at the end. The feeling that we cannot predict what this might imply for art and politics without being pessimistic. Please Philip, keep it up!
@user-lh7mt7zo7l
@user-lh7mt7zo7l Год назад
It just means we'd return to a time before photo and video evidence.
@danisob3633
@danisob3633 Год назад
ye, lie detection needs to get better
@user-lh7mt7zo7l
@user-lh7mt7zo7l Год назад
@Lucas Carvalho I wonder what happens when we make AI generated images of people who don't exist but then someone is born who looks like that haha
@pygmalion8952
@pygmalion8952 Год назад
@@user-lh7mt7zo7l every ai service would be regulated to indicate it is an ai image in the photo's information. tho it is a bit shaky given the fact that you can spoof identification codes sometimes.
@user-lh7mt7zo7l
@user-lh7mt7zo7l Год назад
@@pygmalion8952 yeah regulation wouldn't work because with enough money and power you could make your own A.I.
@existentialselkath1264
@existentialselkath1264 Год назад
New York in unreal engine is genuinely really impressive. It doesn't just look like a game, its got that distinct unreal engine 4 look I can never explain, but it's done it perfectly
@arrowtongue
@arrowtongue Год назад
AI generated images are so great at capturing the feel or vibe of something, because the nature of neural networks is stuff we can't quite describe, it's as scary as it is weirdly comforting we can turn these more abstract feelings into things
@ChrisD__
@ChrisD__ Год назад
I think it's the orange sun like paired with blue everything else, the artist's color grading goals leaking into the actual world lighting. Along with missing shadows here and there and repeated objects and textures. Notice all the fire escapes all over the place. Also the general blurriness of the bounce lighting.
@Strelokos666
@Strelokos666 Год назад
"distinct unreal engine 4 look" what the hell is that suppose to mean?
@ChrisD__
@ChrisD__ Год назад
@@Strelokos666 Ya know... that UE4 look. Orange and teal, TAA, dithering, and every post processing effect under the sun.
@eldarlrd
@eldarlrd Год назад
@@Strelokos666 You haven't played any UE4 game?
@Zoo-Wee-Mama-Sq
@Zoo-Wee-Mama-Sq Год назад
It's been a joy watching your channel branch out from CSGO mapping topics to technology in general, while still bringing the same top notch production.
@distortedjams
@distortedjams Год назад
An interesting test would be to take a real life image, put that into a AI that can transcribe images (Instagram automatically does this). Feed that transcription to one of these AIs and compare the results.
@xouthpaw
@xouthpaw Год назад
And then Instagram won't need any content creators anymore, because you'll be able to log in and receive a 100% AI generated image feed based on your assumed preferences
@IronKurone
@IronKurone Год назад
@@xouthpaw Few years later from today, perhaps my favorite instagram celebrity might not even be human. And that...kinda scary.
@treudden
@treudden Год назад
You can use an init image in disco diffusion which works really good
@Erveon
@Erveon Год назад
@@IronKurone Knowing people have a favorite instagram celebrity is by itself scary enough
@IronKurone
@IronKurone Год назад
@@Erveon Its the future, who know...
@arrowtongue
@arrowtongue Год назад
8:40 your disappointment with the mattresses and stubborn valve please fix made me genuinely burst out laughing, love your sense of humor
@Snowdrift72
@Snowdrift72 Год назад
8:08 is the body the AI has created for itself and chosen to inhabit
@olegmoki
@olegmoki Год назад
If you use DALL-E at 3 am and then turn around... ᅠ
@TilW
@TilW Год назад
I am quite impressed with DALL-E 2, but when closely compared to DALL-E Mini and Midjourney, it still falls behind when it comes to one thing: The inablility to generate Boris Johnson in a bath of beans without violating the TOS.
@mkontent
@mkontent Год назад
This.
@mkontent
@mkontent Год назад
I was honestly shocked that Dall-e mini even knows the faces of popular people. Considering how fuzzy the images are, it will still try its best to recreate Boris Johnson, Ryan Gosling, etc. Literally blew my mind. Like, recognizing faces and the names behind them is something a baby human can do. Little AI that knows Ryan Gosling...
@3n3j0t4
@3n3j0t4 Год назад
@@mkontent dalle mini is literally the only one that allows faces
@3n3j0t4
@3n3j0t4 Год назад
@@mkontent I know you read this you sassy dumbfuck
@explosu
@explosu Год назад
@@mkontent TBF, Boris Johnson already has a face that looks like it was generated by Dall-E Mini so it probably doesn't struggle as much.
@1000_Gibibit
@1000_Gibibit Год назад
Really glad that you managed to get (direct or indirect) access to DALL-E 2. These comparisons are wonderful! And of course you came up with some great prompts for the AI as always. The rate at which AI research advances is actually insane. And the conditions required for this pace, like rapidly improving hardware, are starting to feel like they are straight out of a sci fi story if you think about it. How long before someone accidentally creates an AI that can operate on real life systems that we lose control over? I always thought AI doomsday thinkers were too optimistic about AI. Now I don't know anymore if it's possible for a story like Hyperion to become reality. All bets are off. Oh and all the shorter term consequences relating to reliability of image validity are getting a bit concerning as well of course...
@oldm9228
@oldm9228 Год назад
GitHub copilot is probably an example of a currently active AI that operates on real life systems. It generates context aware code for applications based on requests. The quality of that code is questionable and it could potentially include "hidden intentions" (security risks) just like human written code can.
@HighWarlordJC
@HighWarlordJC Год назад
There's a very real reason many of our brightest minds constantly warn about the dangers of AI.
@amp4105
@amp4105 Год назад
imagine ai generated movies
@amunak_
@amunak_ Год назад
@@oldm9228 Copilot and similar are amazing for generating boilerplate and small chunks of code that you can actually verify yourself. But I have doubts about usage beyond that.
@McDonaldsCalifornia
@McDonaldsCalifornia Год назад
I mean dall-e and gpt and stuff are genuinely impressive but they are far from what we would expect a true AI (or AGI or Super AI or whatever) to look like.
@simian.friends
@simian.friends Год назад
your writing and presentation is particularly great in this video, really enjoyed this, already can tell that I will be rewatching this many times over the coming months
@DeepWeeb
@DeepWeeb Год назад
Petiton to rename the channel to *"3klikspiphlipk"*
@nixel1324
@nixel1324 Год назад
Yes, Dall-e mini (now Craiyon) has a very ai-y feel to it, but I like that. It's like the charm of a retro console. From a technical standpoint it's inferior in every way, but that makes it recognizable, gives it character and makes it endearing. And once people grow up in a world where the higher end stuff is the norm, people like me will probably largely be considered old-fashioned. I don't really care much about modern consoles, and cannot tell apart PS5 and Xbox X footage. But I'll instantly recognize a Wii game, even if running in 4k with texture replacements. Even when you upscale Craiyon results (like with Dall-e Flow), it still has that charm for me. When photo-real ai images become mainstream, I hope people will still appreciate the weirder, less fine-tuned options. I think I will, at least.
@KVVUZRSCHK
@KVVUZRSCHK Год назад
Dall-E Mini is on the left side of the uncanny valley. Astonishingly lifelike yet easily distinguishable as fake.
@IndieLambda
@IndieLambda Год назад
That's when you add "AI generated" at the end of your prompt.
@RaptorShadow
@RaptorShadow Год назад
Someone pointed out that the surreal and disposable quality of the Craiyon images make it perfect for memes. You can quickly and cheaply get a rendering of whatever stupid idea you come up with (like Boris Johnson's Bean Bath Suprise) and get some output. The jank becomes part of the charm.
@juliann149
@juliann149 Год назад
It's crazy watching this video only 9 months later, seeing how much the generators improved already. Would be interesting to re-run the comparison on the current versions, especially Midjourney advanced a lot iirc.
@Tofuey
@Tofuey Месяц назад
Even further a year from this comment
@mauricepouly
@mauricepouly Год назад
i adore your videos and the flair you bring into them. i enjoyed this one a lot too and it made me laugh which is a feat on its own. thank you phillip keep on doing what you do
@hisshame
@hisshame Год назад
Thank you for sharing the process with us!
@digitalrockets9702
@digitalrockets9702 Год назад
That's so wild how I stumbled across your channel. While I was generating my own thing on the midjourney discord, I distinctly remember seeing your Martian base castle iterations pass by in the queue as well. So interesting how everyone can just watch each other's image generation take place at the same time. like watching parallel worlds unfold.
@emperorpalpatine1469
@emperorpalpatine1469 Год назад
Mate I'm so glad you're still making videos like this, you're probably my favorite chap on RU-vid. You taught me how to play counter strike and you got me into technology from a young age, thanks a lot Mr Phillip :)
@llave8662
@llave8662 Год назад
DALL-E can't generate words to avoid falsifications, similarly to the reason why it does not allow faces. Great video!
@LazyBuddyBan
@LazyBuddyBan Год назад
thats explains it. but also won't matter, since likely in 5 years we will get it without restrictions.
@manfail7469
@manfail7469 Год назад
@@LazyBuddyBan christ, can you imagine how much the world will change when stuff like dall-e 2 go unrestricted?
@JustSayin24
@JustSayin24 Год назад
Actually the original research paper for DALL-E 2 states that text rendering is a known limitation of the model. Specifically, the "embedding does not precisely encode spelling information of rendered text" - in other words, the model isn't trained at a high-enough precision to properply represent the intracacies of charecter shapes and grammatical rules.
@tissuepaper9962
@tissuepaper9962 Год назад
@@JustSayin24 I imagine they are already working on making the next model recognize text in the training data, transcribe it, and run it through a separate NLP model so that the image generator can understand grammar and spelling and stuff.
@cem_kaya
@cem_kaya Год назад
​@@tissuepaper9962 there is no need to do such convoluted stuff to get photorealistic text generation. scaling up the model works fine
@devindykstra
@devindykstra Год назад
Of all the things to defeat Dalle 2, I would never have expected a mattress leaning against a wall.
@kruchji
@kruchji Год назад
I love how you immediately answer any question that I think of while watching. Great video!
@vekst
@vekst Год назад
Some of those film imitation ones are amazing, and convinced me to hoin the DALL-E 2 waitlist to try some for myself. Great vid as always Philip!
@raminasadollahzadeh328
@raminasadollahzadeh328 Год назад
I am starting to get back to your channels. Been gone for about 2 years and now that I'm back I have to say there is a style in your videos which is very rare and unique. People say CSGO utubers are dying but you are proving them wrong. I am happy for you and for my self to find my new old fav channel.
@mgetommy
@mgetommy Год назад
So cool, well done Philip . I love your tendency to see something interesting and tweak with it and show us the results
@HELLF1RE9
@HELLF1RE9 Год назад
8:06 that is unbelievably unnerving
@mixchief
@mixchief Год назад
1:00 Hahahahaha! Beautiful piece. Like the subtle touch with the blue wig covering half his hair. And 1:05 gets even better. That a turd floating around in the bean soup?
@maximiliankegley-oyola928
@maximiliankegley-oyola928 Год назад
These are all fascinating. I love seeing the differences between the AIs. Always love your videos!
@Revan-kq7ih
@Revan-kq7ih Год назад
Some of these pictures made laugh really hard. Great video, we need more of these!
@luigimaster111
@luigimaster111 Год назад
This already has the potential to be an immense asset for artists, I for example dabble at making and animating 3d models and on a frequent basis struggle to find/make good textures to use on my models, now it's pretty much at a point where I could just ask an AI for what I need, do a bit of tweaks, then plop it in. Now we just need an AI for generating things like normals maps, as the existing automatic tools I've found to be a bit lackluster. I also struggle to visualize things, that is to say I can't visualize things at all to the point where even my dreams are an entirely auditory experience, so when making anything I have to make heavy use of references to get anywhere. Tools like this of course make getting specific reference material easier. With so much progress happening in such a relatively short time span though... Well I think stuff like this will soon threaten the jobs of a lot of people. I for example am suggesting to use them to fill the roll of texture and concept artist for my small scale personal projects, at what point will large studios decide to do something similar?
@artizard
@artizard Год назад
i was hoping you were going to make a video about dall E 2, nice video! You should do even more ai related videos, i find them really fascinating
@deKxiAU
@deKxiAU Год назад
Just a tip for 'photorealism' with these models: put camera / photography specs like F-stop, iso and lens length - works best with DALLE 2 Edit: ah I see you've done that in the later prompts, ignore the above then :P Also worth noting none of the generation methods actually merge images from Google, they just had watermarked images in the dataset. I realise you probably don't think that it actually does just google some images given what you said, and it might seem pedantic - but it's a proper distinction (and the 'just slapping images from googling together' myth is a very common for all AI generative art right now), for those who don't know: the model's learned that that particular image is likely to have a watermark from what its seen in its training dataset and so it's synthesised it. It's not actually searching anything on any search engine or anything like that, it's just a matter of the dataset not being cleaned of any watermarked content Great video though Philip :) Edit: also for Midjourney specifically, there's some additional background style modifiers you can disable that would be somewhat influencing your result out of the box for the cartoon ones and making them less accurate to the prompt, forgot what the arg is as I'm on mobile watching this but it's somewhere in the FAQ I believe - but this is why you always get a vignette, a similar colour palette, amongst others behaviour across every Midjourney prompt
@huttyblue
@huttyblue Год назад
What was a watermarked image doing in its data set if it wasn't sourced from scraping the web though? It may not be specifically from google but the concept of it just learning from what was able to be searched up on the internet is the same.
@deKxiAU
@deKxiAU Год назад
​@@huttyblue I didn't say it wasn't due to web scraping, it absolutely is. I'm saying it's not googling/searching online at inference stage (the stage where people can actually interact with the AI in the way you see in the video), and in fact the AI never touches the internet (outside of it getting hosted online to be accessed by yourself, or it being trained using a GPU cloud farm somewhere in the first place). It's a large fundamentally different procedure with very different outcomes. Since most people here probably aren't familiar with neural net training I'll elaborate a bit: CLIP was trained on web scraped images as outlined above (CLIP being the model under the hood of most AI generative apps/notebooks and of Midjourney too), but it's nowhere near the same as a program searching up your prompt for close images online that match and then splicing those together - it's not a glorified Pinterest board. The dataset is static from the date of when it was scraped and published. It's then used as training material for the basis of the AI's generative ability - you won't find things posted after the date of the dataset in it's generative vocabulary for example. Naturally, a poor dataset can lead to poor results and watermarks are an obvious poor side effect of web scraping, but to conflate it with 'searching online' gives the impression that it's simply reverse searching for your images and slapping them together which leads to people believing that it is actively searching and essentially 'cheating' - like someone looking up results for a test right? Whereas in actuality it genuinely generates the images based on what it learned from 'studying' the dataset and associating different labels with what it thinks is relevant, as if it spent it's time studying wikipedia articles instead of the sources wikipedia lists, etc. CLIP has stupidly learned that stock image watermarks are common enough across it's whole dataset that they are worth adding to some images sometimes even when not directly prompted for it, because it had enough images to train on that had watermarks that it associated the concept of watermarks with that sort of image in its latent space. But it's the concepts themselves that it has learned, not direct image portions and mashing them together. DALL-E 2 has this same issue but the dataset was far more curated, it's fairly difficult to get a blatant example. DALL-E Mini (now CrAIyon) also suffers from this but the quality is bad enough that you'll be hard pressed to even recognise it's a watermark and not just random jibberish text. Most models at the moment are trained with the LAION dataset (among a few more) which has a whole host of web scraped content (including graphic porn and all sorts of NSFW images - these usually get taken out manually by the big companies models), but until there are open sourced datasets that don't have to rely on web scraping to get the sheer number of images training a model requires (several hundred million to billions), stuff like watermarks and weird quirks are just part of the parcel - that said, web scraping is also why it can make such hilarious memes because the highly curated datasets (like the one in DALL-E 2) remove large chunks of the image base and sort of gut the models ability to accurately reach a prompt in the process. TLDR: Its the difference between studying for a test before the day, or actively searching online during your test. Hope that helped illuminate the differences! Enjoy your day :)
@tissuepaper9962
@tissuepaper9962 Год назад
@@deKxiAU I disagree that there's much of a difference. The claim that it's "just slapping images together" is basically pointing out that the system doesn't know anything about *why* certain features exist in images, it just knows that they *do*. AI at this point are still just advanced statistical aggregators, most lack the kind of logic that would allow them to generate images with details that make sense as opposed to just looking right at a glance. Philip isn't saying that the system literally merges images from Google at the time of inference, it seems to me like a subtle statement about "learning" vs. "regurgitating" and what should actually be called "intelligence".
@deKxiAU
@deKxiAU Год назад
@@tissuepaper9962 there is a significant difference. If Philip meant that 'it doesn't know why features exist' he should have just said that. Learning the 'wrong' details doesn't make it 'regurgitation' any more than learning the right details would, and it falling apart under scrutiny is largely due to the limited resolution of the training data (typically 256x256 or 512x512, DALL-E 2 starts at 64x64 with additional diffusion networks trained on upscaling it incrementally) combined with the limited number of parameters the model contains which leads to it having to combine concepts into the same latent dimensions and differentiate between them poorly as a result. I'm not sure what you could disagree with really, like I said at the bottom - it's the difference between studying for a test or looking up the answers during it, entirely different implications can be drawn from systems that do either of those. The former relies on prelearning concepts and identifying key relationships between them, the latter can pick new images as they pop up on the internet and doesn't have any understanding of the relationship between concepts at all. One is learning conceptual relationships, the other is a pinterest board with a fancy text input. I'm not saying it's not statistical aggregation, I'm just saying it's not ripping images off the internet and splicing them together like some Frankenstein creation, and that it *has* learned within the weights of its millions to billions of parameters that there is an association between watermarks and those types of images - which is actually true, in the dataset it was trained on there was enough watermarks for it to recognise the concept across them and learn about it the same way it has for every other concept it recognised; like trees and bushes belonging in a forest, stock watermarks belong on stock-looking images. Removing the watermarks from the dataset would solve that specific issue, but wouldn't change anything about how its fundamentally working, it would just give better results because it's an algorithm that aims to be able to create images that *could* have been from its dataset without actually recreating any image from it (that would be what's called overfitting, which we dont see in these models), its task is quite literally to map the entire range of possible images in its dataset and to abstract whatever relationships it can to condense it into its embedded parameter weights, and so it would be a failure if it didn't have watermarks when there are so many in the dataset. Make sense? Intelligence has nothing to do with it, different conversation entirely. Not arguing it's sentient or that it understands in a way that human brains do, (obviously, the way it understands and learns isn't as complex as humans and it doesn't have an understanding of *why* these things exist together, just that they do, because the why wasn't part of the training data - its simply condensing image concept relationships to an extremely large matrix of numbers), just that people shouldn't propagate a myth because "it's close enough" when it actually gives a false impression of what these models can do and how they work, and what it means for the world; different behaviour, different results, different legal implications, different world outcomes and use-cases. I'm only hoping to help correct the record as I'm a huge fan of Philips content - not wanting to knock the video, overall it's very good and knowledgeable and at the incredibly high standard Philip always provides for his videos - just that particular line (which he said twice) suggests hes either a bit misinformed on the topic (which is fine, everyone's misinformed on something and it shouldn't be taken personally if it's corrected) or that he wasn't quite clear on what he meant (also fine as he possibly wasn't aware of how what he said could be interpreted)
@tissuepaper9962
@tissuepaper9962 Год назад
@@deKxiAU You have your interpretation, I have mine. You can carefully defend the model by explaining the limitations, that doesn't change my opinion. I think it's a perfectly acceptable simplification made for brevity, something you appear to hold in little regard. PS: You say "intelligence" is a different discussion, did you forget what "AI" stands for?
@akkkarinn
@akkkarinn Год назад
I want MORE, we need a second part to this video!
@Lulzalex
@Lulzalex Год назад
First the silent zoom in on the horse shaped entity followed by doing the same to the generated Chucky-doll all completely threw me off LMAO. I kept checking my back for the remainder of the video and could not relax like I usually do...
@yom35
@yom35 Год назад
Amazing video as always!!
@ozmog6458
@ozmog6458 Год назад
Hey, thanks for making these videos.
@dominikrohel2546
@dominikrohel2546 Год назад
I appreciate your passion in making these intersting videos. I wonder what the technology will be like in 5 or 10 years. I guess i’ll have to wait for future philip to cover it
@Sgt_Recka
@Sgt_Recka Год назад
I know you said that people don’t watch this kind of content from you. But I just wanted to say that I love it! AI is so interesting, and not many people on RU-vid are showing what you are showing. I’m here for all your content, from all 3 channels
@BigCheeseEvie
@BigCheeseEvie Год назад
Splendid content, really entertaining. Keep em coming!
@h930hec
@h930hec Год назад
Fantastic video Philip. Very much enjoying the AI content!
@artemisDev
@artemisDev Год назад
the new Turing test: "Draw a mattress leaning up against wall".
@broomguy7
@broomguy7 Год назад
Another great video from 36PKL!
@lazz4205
@lazz4205 Год назад
Dall-e mini is an amazing tool to fiddle around with, i find it really excels at abstract depictions of things - the style "album cover" can come up with some pretty cool stuff when paired up with good prompts
@HSE_VO
@HSE_VO Год назад
I adore your AI videos. Please keep them going!
@CrazyKosai
@CrazyKosai Год назад
more shenanigans with DALL-E 2 plz
@seto007
@seto007 Год назад
Hey Philip, I recently got access to both DALL-E 2 and Midjourney, and so I wanted to share a bit of my perspective on the strengths and weaknesses of both. While DALL-E 2 is certainly better at generating the initial image at a higher fidelity and with more stylization based on the description provided, I actually think that Midjourney succeeds far more at creating a "final image" than DALL-E 2 does. The reason for this is that the subsequent variations of an image that you can generate with DALL-E 2 often deviate significantly from the original description, to the point where it often feels as though the AI is trying to guess at what the original description you used was based on the image it's making variations from, and because of this it often feels like the AI subsequently gets confused and creates more abstract renditions than what you might have intended. Midjourney doesn't seem to have this issue. Subsequent variations seem to stick to both the original description and intent behind the image it's creating variations of, and because of this it feels as though subsequent generations look much closer to the original intent of the person describing the image to be generated. Beyond this, it feels as though DALL-E 2 has some issues with understanding things like perspective in all but the most simple of circumstances. If you were to ask it to generate an image viewed from the side, for example, it will often give you an image viewed from a diagonal downwards angle, as opposed to a true sideshot like what you would see in something like a Shutterstock photo. Midjourney does not have this issue in most circumstances; it seems to understand that you want to view the object being described from a side-facing angle. I think both models have their strengths and weaknesses, depending on the use case; since I am primarily interested in using these AIs to speed up the art process for a cyberpunk video game I am working on, I like using DALL-E 2 to generate stylized concept art that gets across the themes I am going for, whilst I prefer using Midjourney to generate more technical images of hypothetical in-game objects to use as reference.
@Gheno
@Gheno Год назад
Perfect timing, just got done draining my quest's battery and I need something to watch while eating. Thanks, unc phil.
@TheKrzysiek
@TheKrzysiek Год назад
While others are worried about using this for more malicious stuff, I'm more excited about how much cool new content we can get from this. Want a specific image for a video, wallpaper, or a meme? Put it in AI I especially wonder if it will ever be used for things like concept art, book covers, character portraits etc.
@FredMoin
@FredMoin Год назад
I thought some of your upscaling videos were interesting, but this is just amazing. Thanks for the work you put in to show us what AI is capable of. Do you know if DALL-E 2 is made for that resolution and could be adjusted to put out more realistic images with more time/processing? Anyway, i still have a hard time to accept that a "program" can interpret text into relating images at all.
@loetwiek
@loetwiek Год назад
i love the ai things on your channel keep em coming
@arandompotat0
@arandompotat0 Год назад
Love the AI image generation videos. Hope you can continue making these videos, since the tech will constantly improve. As a digital artist, your experimentations videos testing the limits of AI are so captivating. Will I have a job in a few years? probably not in concept art. Not anymore, that's for sure 😂😅
@Lohmeier54
@Lohmeier54 Год назад
I remember your old videos on ai face improvement. I never expected ai to get here until i was at least 30. I'm not even 20, this is incredible
@flyinggoatman
@flyinggoatman Год назад
I'm so glad I got access to both.
@JohnDoe-sw2nc
@JohnDoe-sw2nc Год назад
DALL-E 2 is scary good
@arcadianpunk
@arcadianpunk Год назад
Let the battle begin
@Senkiowa
@Senkiowa Год назад
8:13 The mattress things is interesting as generally "clutter" is one thing I always tought make realistic renders distinguishable from real life. There is so much unique stuff in the word that people just leave lying around that someone working on an appeasing looking 3D scene will not be able or want to include or will only include it in a way that makes it identifiable. While in reality if you look at a photograh you probably going to see blobs of things that can't be identified.
@mennonis
@mennonis Год назад
Actually watching this to keep up with the progress, as I feel it will be important
@adan7949
@adan7949 Год назад
It's actually a bit scary how convincing some of the images are, I'm glad those things are hard to get your hands on
@declanlambert1089
@declanlambert1089 Год назад
not for long
@BombBird11
@BombBird11 Год назад
@@happygofishing Dangerous little man, now aren't we? lol
@gigabooga
@gigabooga Год назад
@@happygofishing Yeah but no one asked
@spiderjerusalem8505
@spiderjerusalem8505 Год назад
@@happygofishing, true
@ShawnFumo
@ShawnFumo Год назад
MidJourney has been letting many people in lately. Even DALL-E 2 said they let in 10k people in a week recently and had a survey on pricing models. It won’t be long.
@MarkSulekTalk
@MarkSulekTalk Год назад
For your interest, I've seen an article where a photographer integrated an image he took in DALL-E 2, which was blurry and out of focus, and writted "Ladybug on a leaf, focus stacked high resolution macro photograph". The image recovered details and focus and became tack sharp, which was impressive ! You should try to do that !
@tedstriker2000
@tedstriker2000 Год назад
looking forward to these doing animations '')
@Diie89
@Diie89 Год назад
That final image with the men in purple jumpsuits singing in to a microphone is quite horrifyingly realistic. Even when zooming in and trying to spot things that might be off, I still struggle to find stuff.
@KrynexYT
@KrynexYT Год назад
As Károly from Two-Minute Papers always says, imagine the improvement two papers down the line. DALL-E 4 will probably make graphic designers etc. largely redundant.
@bluebell2334
@bluebell2334 Год назад
I love Karoly's style of presenting something. Each video exceeds my expectations.
@s-zz
@s-zz Год назад
The irony of it all, is the fact that a lot of the same AI designers are also working on AI that can code. And will eventually cause them to become obsolete too. Seriously, look up coding with AI, there's a lot of info on it already.
@rene-of3sc
@rene-of3sc Год назад
​@@s-zz Meh, Copilot for example is useful to generate easy or repetitive functions but no matter what, a human would need to say what needs to be generated and see if the generated code is correct. I would assume in the future AI will be used as a productivity tool by programmers, but not replace them.
@AlphaGarg
@AlphaGarg Год назад
@@rene-of3sc This. I hate this whole "[job] will become redundant!" falsity that people have for some reason hung onto. Did Photoshop make photographers' jobs redundant? No! Did node-based programming like Unreal's blueprints make programmers' jobs redundant? No! Neural networks like DALL-E, Jukebox, etc. are tools that'll be used by the people that know the most about these things - artists. Sure, any old schmuck might be able to generate an image based on a prompt. But they aren't going to be able to do it the same way an artist will. Artbreeder has existed for a while now, yet outside of artist circles, I haven't seen that much use of it. Same will happen to these once they get normalised and become accessible.
@trallakid
@trallakid Год назад
i don't think it will make all graphic designers redundant, just the ones stuck in the past. with all professions, the technology is constantly changing so any good graphic designers would ideally use these types of technology as another tool in the toolbelt. As a graphic designer myself I can 100% see this technology being great for idea generation and coming up with some ideas from prompts, but I don't think it will ever be able to fully replace a human (although mark my words I might regret going down this career path in a few years lol)
@christophernoneya4635
@christophernoneya4635 Год назад
I think my favourite use for Dall-e mini is generating abstract art. It does this really well as it becomes easy to ignore say the smudging on the sides of an eye. It really does seem to struggle specifically with the human form, as one part bleeds into another.
@harrymalm
@harrymalm Год назад
I was expecting this to be on about the same level as the competitions you made between the CS bots, but I was wrong... Some of the images generated by DALL-E 2 could definitely fool me, and I think you should make a video where the viewer guesses which images are real and which ones have been generated.
@sanderbos4243
@sanderbos4243 Год назад
Loved this video!
@bjk0norway0bjk
@bjk0norway0bjk Год назад
really enjoyed this video :D
@pastfuturizednow7907
@pastfuturizednow7907 Год назад
thank you for your work
@QuestioningYourSanity
@QuestioningYourSanity Год назад
This is beyond fascinating. If I were making a movie or video game, I would use this to expand my idea of whats possible.
@jetex1911
@jetex1911 Год назад
Dall-e might not be able to do full images well, but I've definitely been having fun getting it to generate ideas for art thumbnails I could bring to life
@PCubiles
@PCubiles Год назад
Dall-E 3 (or at least something published at around the same time) will most likely be generating videos, there's already small examples that can do that from 1 initial image, but if you can connect that to a firstly generated image we could get it in a few years, or even in just one
@MattVidPro
@MattVidPro Год назад
great video! I've been making a plethora of videos discussing and testing this technology lately, and man is it moving FAST. Every few days I hear something new....
@iulic9833
@iulic9833 Год назад
I know, can't wait for DALLE 2 to get released to the public, if it ever will. Also got some good results when upscaling the images too, they have some artifacts but its still mind boggling how an AI can create stuff as this. ru-vid.com/video/%D0%B2%D0%B8%D0%B4%D0%B5%D0%BE-a1J3ToFUAX8.html&ab_channel=69fff
@godofzombi
@godofzombi Год назад
I've found Dall E mini is really good at Art Deco posters, especially if you stick to natural landscapes. H.R. Giger also gives decent results, altough not the best quality. And mini's tendency to mangle faces makes some drawings almost look like the works of Francis Bacon.
@lukasg4807
@lukasg4807 Год назад
TBH I'm more impressed with the ability to understand what you're asking for than the image generation itself
@vectornine
@vectornine Год назад
The faces at the end are so good it's crazy
@amnesicpachyderm
@amnesicpachyderm Год назад
I'm loving these AI videos. It's an exciting and worrying time, and it feels like we're on the precipice of some historic developments. Hopefully good ones. But I guess we'll see either way.
@beanbandit495
@beanbandit495 Год назад
The ending completely blew my mind!
@seruppo4219
@seruppo4219 Год назад
Thank you for this video klik, have a wonderful day or and night.
@boofkoosher2631
@boofkoosher2631 Год назад
I was very shocked with dall-e 2 results. They were scaring-ly accurate and very detailed.
@Periwinkleaccount
@Periwinkleaccount Год назад
Scarily.
@boofkoosher2631
@boofkoosher2631 Год назад
Thank you dear sir for providing an appropriate word to fixate my lingua franca
@kriterer
@kriterer Год назад
Seeing Boris emerge from the beans at 0:55 is immaculate
@lucasliboiron7342
@lucasliboiron7342 Год назад
Dude great video as always. Love your videos. Love this topic. Dont get too deep mentally in AI taking over though bro! hahaha! I see youve been INTO this topic. Anyways, i Love it and keep the videos coming. Humans #1
@kasplay7275
@kasplay7275 Год назад
the pics with the text in them is just like when u try to read in your dream
@keenban
@keenban Год назад
I just got access to Dall-E 2 the other day, and I have been playing around with it. Honestly, it is quite crazy what it can do. I wonder how it would be if it were unrestricted.
@BombBird11
@BombBird11 Год назад
*C H A O S .* Just pure, utter chaos....💀
@LucUltra
@LucUltra Год назад
That was very interesting and entertaining. Excellent prompts too! I hope they expand access to DALL-E 2 soon. Do you feel like you could use these generators for your creative process? If so, how?
@KinnetMontana
@KinnetMontana Год назад
A csgo map made completely out of ai concept art would be interesting
@mattd1466
@mattd1466 Год назад
I'm not sure you're aware of how good you are at presenting and making topics interesting, like I still watch your csgo videos even though I haven't been playing the game for years because they're still enjoyable to watch.
@mattd1466
@mattd1466 Год назад
@@2kliksphilip oh totally! at the end of the day I prefer my Philip kliked twice over thrice.
@MikeKleinsteuber
@MikeKleinsteuber Год назад
Nicely done.
@CozMyN
@CozMyN Год назад
I lost it when you said "Valve, please fix" :)))))
@GamerReality
@GamerReality Год назад
What music do you use in you videos? So enjoyable to listen to while you're talking!
@MrRobotrax
@MrRobotrax Год назад
I'd really love to see dall e 2 make images of the backrooms. It feels like the perfect prompt for AI, since the images it generates already have a somewhat dream-like aura to them.
@gtPacheko
@gtPacheko Год назад
Great videl from 26PKL!
@RatedX29
@RatedX29 Год назад
great video 36pkl like always
@WatamelonUberSheep
@WatamelonUberSheep Год назад
Son: Mom I want an EVA Mom: We already have an EVA at home sweetie Eva at home: 1:54
@laprueba6521
@laprueba6521 Год назад
1:40 comically small bed
@DaxyGamer
@DaxyGamer Год назад
2kliksphilip's MAN-SIZED bed
@markoforfun
@markoforfun Год назад
Btw , i liked the first title+tumbnail better , even set it as watch later to check a look.But good video +1
@luna010
@luna010 Год назад
tbf, midjourney’s first result was definitely the most interesting, and I think it fulfilled the prompt. the dalle2 results look like shitty google images clipart.
@ipixz3
@ipixz3 Год назад
Wouldn't it make more sense for Concept Artists to be considered obsolete instead?
@luna010
@luna010 Год назад
I feel like the more realistic AI generated images become, the more people will appreciate how cool “bad” AI generated images can be. The novelty of photorealistic images being AI generated will wear off once it’s commonplace, but images that don’t look real will always be at least a little bit interesting.
@SBImNotWritingMyNameHere
@SBImNotWritingMyNameHere Год назад
put timestamp so more people get what youre talking about pls
@Veptis
@Veptis Год назад
And just a few months later there is even more models. Dall-E 2 is great because it can also take an image, prompt and mask as input. The midjourney training set does include "art" and doesn't do photos well. Dall-E 1(and it's mini clone) is using patches, Dall-E 2 is using diffusion... Which explains the difference in results.
@Chreeeis
@Chreeeis Год назад
That text rendering was spine crawling
@Nyllsor
@Nyllsor Год назад
Very informative!
@Failzz8
@Failzz8 Год назад
Dude the matress part fucking killed me, why is this so funny holy shit lol
@AlessandroBluesBreaker
@AlessandroBluesBreaker Год назад
The vertical matress thing is crazy i went through a whole phase where i disassembled my bed for more space
@phntm5700
@phntm5700 Год назад
this implies so much about the future
Далее
Portal RTX
11:44
Просмотров 547 тыс.
I'm a NERD??
4:51
Просмотров 47 тыс.
VLOG: ПОДАРИЛА МАШИНУ РОДИТЕЛЯМ
27:46
A.I art is really terrifying!
15:09
Просмотров 234 тыс.
How Much Power Does Everything Use?
12:07
Просмотров 116 тыс.
de_prime - How Far from the Source can it get?
11:07
Просмотров 515 тыс.
HL2 Citadel Explosion recreated in Teardown
0:53
Просмотров 20 тыс.
The Dall-E AI Has Gone Too Far...
9:58
Просмотров 559 тыс.
Segmented Displays
13:55
Просмотров 4,9 млн
Look what you made me do
5:21
Просмотров 52 тыс.
Can you tell what's real? - AI Generated Videos
7:16
Просмотров 313 тыс.
VR Guide 2022 and Beyond
8:06
Просмотров 147 тыс.
Body Symbol Game With Sonic And Shadow
0:23
Просмотров 3,9 млн